[ 527.327108] env[65107]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=65107) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 527.327348] env[65107]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=65107) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 527.328507] env[65107]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=65107) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 527.328866] env[65107]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 527.444576] env[65107]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=65107) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 527.453879] env[65107]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=65107) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 527.499274] env[65107]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 528.059029] env[65107]: INFO nova.virt.driver [None req-46e5c6d4-c5ef-4f16-9c03-0b0205e41dc6 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 528.128413] env[65107]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 528.128590] env[65107]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 528.128684] env[65107]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=65107) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 531.347443] env[65107]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-3c2f5d9c-cdca-4237-891d-4512ee38ad2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.364295] env[65107]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=65107) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 531.364522] env[65107]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-188c2a4d-4e5a-4880-894b-3aa1ec26098f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.397888] env[65107]: INFO oslo_vmware.api [-] Successfully established new session; session ID is c6fbd. [ 531.398046] env[65107]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.269s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 531.398739] env[65107]: INFO nova.virt.vmwareapi.driver [None req-46e5c6d4-c5ef-4f16-9c03-0b0205e41dc6 None None] VMware vCenter version: 7.0.3 [ 531.402137] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1267c496-afe0-4017-bb87-33c315ad9fda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.420015] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ad7802-8e10-49bb-949b-1d24c55539a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.426273] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e302d9b-edea-4d62-ac32-400fbb64dcb4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.433029] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363238c2-e78b-46e2-8c60-df06f520edf0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.446264] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25f6707-8217-45bf-a6f2-6c8917b0abe6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.452614] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96051d76-f097-4749-94f4-12f05213c32a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.483686] env[65107]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-361fbd65-6cc3-428d-b528-0d61c500904c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.490349] env[65107]: DEBUG nova.virt.vmwareapi.driver [None req-46e5c6d4-c5ef-4f16-9c03-0b0205e41dc6 None None] Extension org.openstack.compute already exists. {{(pid=65107) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 531.493037] env[65107]: INFO nova.compute.provider_config [None req-46e5c6d4-c5ef-4f16-9c03-0b0205e41dc6 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 531.995845] env[65107]: DEBUG nova.context [None req-46e5c6d4-c5ef-4f16-9c03-0b0205e41dc6 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),45ed7fb8-8e0d-4f49-8dae-53b2ef6d82c5(cell1) {{(pid=65107) load_cells /opt/stack/nova/nova/context.py:472}} [ 531.996182] env[65107]: INFO nova.utils [None req-46e5c6d4-c5ef-4f16-9c03-0b0205e41dc6 None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 531.998137] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 531.998351] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 531.999043] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 531.999491] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Acquiring lock "45ed7fb8-8e0d-4f49-8dae-53b2ef6d82c5" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 531.999672] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Lock "45ed7fb8-8e0d-4f49-8dae-53b2ef6d82c5" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 532.000689] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Lock "45ed7fb8-8e0d-4f49-8dae-53b2ef6d82c5" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 532.020868] env[65107]: INFO dbcounter [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Registered counter for database nova_cell0 [ 532.029794] env[65107]: INFO dbcounter [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Registered counter for database nova_cell1 [ 532.033125] env[65107]: DEBUG oslo_db.sqlalchemy.engines [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65107) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 532.033524] env[65107]: DEBUG oslo_db.sqlalchemy.engines [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65107) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 532.038475] env[65107]: ERROR nova.db.main.api [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 532.038475] env[65107]: func(*args, **kwargs) [ 532.038475] env[65107]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 532.038475] env[65107]: self.work.run() [ 532.038475] env[65107]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 532.038475] env[65107]: result = self.fn(*self.args, **self.kwargs) [ 532.038475] env[65107]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 532.038475] env[65107]: return func(*args, **kwargs) [ 532.038475] env[65107]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 532.038475] env[65107]: result = fn(*args, **kwargs) [ 532.038475] env[65107]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 532.038475] env[65107]: return f(*args, **kwargs) [ 532.038475] env[65107]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 532.038475] env[65107]: return db.service_get_minimum_version(context, binaries) [ 532.038475] env[65107]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 532.038475] env[65107]: _check_db_access() [ 532.038475] env[65107]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 532.038475] env[65107]: stacktrace = ''.join(traceback.format_stack()) [ 532.038475] env[65107]: [ 532.039629] env[65107]: ERROR nova.db.main.api [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 532.039629] env[65107]: func(*args, **kwargs) [ 532.039629] env[65107]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 532.039629] env[65107]: self.work.run() [ 532.039629] env[65107]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 532.039629] env[65107]: result = self.fn(*self.args, **self.kwargs) [ 532.039629] env[65107]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 532.039629] env[65107]: return func(*args, **kwargs) [ 532.039629] env[65107]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 532.039629] env[65107]: result = fn(*args, **kwargs) [ 532.039629] env[65107]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 532.039629] env[65107]: return f(*args, **kwargs) [ 532.039629] env[65107]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 532.039629] env[65107]: return db.service_get_minimum_version(context, binaries) [ 532.039629] env[65107]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 532.039629] env[65107]: _check_db_access() [ 532.039629] env[65107]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 532.039629] env[65107]: stacktrace = ''.join(traceback.format_stack()) [ 532.039629] env[65107]: [ 532.040127] env[65107]: WARNING nova.objects.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 532.040272] env[65107]: WARNING nova.objects.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Failed to get minimum service version for cell 45ed7fb8-8e0d-4f49-8dae-53b2ef6d82c5 [ 532.040695] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Acquiring lock "singleton_lock" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.040847] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Acquired lock "singleton_lock" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 532.041095] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Releasing lock "singleton_lock" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 532.041414] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Full set of CONF: {{(pid=65107) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:275}} [ 532.041549] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ******************************************************************************** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 532.041672] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] Configuration options gathered from: {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 532.041799] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 532.041985] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 532.042125] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ================================================================================ {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 532.042328] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] allow_resize_to_same_host = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.042495] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] arq_binding_timeout = 300 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.042616] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] backdoor_port = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.042736] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] backdoor_socket = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.042890] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] block_device_allocate_retries = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.043055] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] block_device_allocate_retries_interval = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.043230] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cell_worker_thread_pool_size = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.043392] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cert = self.pem {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.043554] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.043714] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute_monitors = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.043873] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] config_dir = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044045] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] config_drive_format = iso9660 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044176] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044335] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] config_source = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044495] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] console_host = devstack {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044651] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] control_exchange = nova {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044806] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cpu_allocation_ratio = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.044955] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] daemon = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.045128] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] debug = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.045302] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_access_ip_network_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.045478] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_availability_zone = nova {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.045624] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_ephemeral_format = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.045774] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_green_pool_size = 1000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.046012] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.046194] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_schedule_zone = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.046417] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] default_thread_pool_size = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.046583] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] disk_allocation_ratio = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.046742] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] enable_new_services = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.046898] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] flat_injected = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.047064] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] force_config_drive = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.047220] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] force_raw_images = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.047388] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] graceful_shutdown_timeout = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.047574] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] heal_instance_info_cache_interval = -1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.047815] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] host = cpu-1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.047989] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.048164] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] initial_disk_allocation_ratio = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.048326] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] initial_ram_allocation_ratio = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.048538] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.048696] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_build_timeout = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.048848] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_delete_interval = 300 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049014] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_format = [instance: %(uuid)s] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049190] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_name_template = instance-%08x {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049348] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_usage_audit = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049511] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_usage_audit_period = month {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049667] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049824] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] instances_path = /opt/stack/data/nova/instances {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.049978] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] internal_service_availability_zone = internal {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.050144] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] key = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.050297] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] live_migration_retry_count = 30 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.050459] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_color = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.050614] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_config_append = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.050769] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.050922] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_dir = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051088] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_file = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051216] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_options = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051373] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_rotate_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051536] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_rotate_interval_type = days {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051695] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] log_rotation_type = none {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051818] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.051938] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.052114] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.052277] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.052405] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.052559] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] long_rpc_timeout = 1800 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.052711] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] max_concurrent_builds = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.052859] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] max_concurrent_live_migrations = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053014] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] max_concurrent_snapshots = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053198] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] max_local_block_devices = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053360] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] max_logfile_count = 30 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053511] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] max_logfile_size_mb = 200 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053660] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] maximum_instance_delete_attempts = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053810] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] migrate_max_retries = -1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.053966] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] mkisofs_cmd = genisoimage {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.054183] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] my_block_storage_ip = 10.180.1.21 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.054311] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] my_ip = 10.180.1.21 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.054513] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.054670] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] network_allocate_retries = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.054842] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.055011] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] osapi_compute_unique_server_name_scope = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.055177] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] password_length = 12 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.055356] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] periodic_enable = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.055518] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] periodic_fuzzy_delay = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.055684] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] pointer_model = usbtablet {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.055843] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] preallocate_images = none {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.056007] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] publish_errors = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.056140] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] pybasedir = /opt/stack/nova {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.056365] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ram_allocation_ratio = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.056573] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] rate_limit_burst = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.056740] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] rate_limit_except_level = CRITICAL {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.056892] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] rate_limit_interval = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057057] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reboot_timeout = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057214] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reclaim_instance_interval = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057366] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] record = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057528] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reimage_timeout_per_gb = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057688] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] report_interval = 120 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057841] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] rescue_timeout = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.057993] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reserved_host_cpus = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.058160] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reserved_host_disk_mb = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.058314] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reserved_host_memory_mb = 512 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.058470] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] reserved_huge_pages = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.058623] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] resize_confirm_window = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.058774] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] resize_fs_using_block_device = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.058926] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] resume_guests_state_on_host_boot = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.059103] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.059265] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] rpc_response_timeout = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.059612] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] run_external_periodic_tasks = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.059808] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] running_deleted_instance_action = reap {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.059967] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] running_deleted_instance_poll_interval = 1800 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.060136] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] running_deleted_instance_timeout = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.060292] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler_instance_sync_interval = 120 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.060454] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_down_time = 720 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.060618] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] servicegroup_driver = db {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.060768] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] shell_completion = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.060919] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] shelved_offload_time = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.061082] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] shelved_poll_interval = 3600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.061248] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] shutdown_timeout = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.061405] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] source_is_ipv6 = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.061557] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ssl_only = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.061816] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.061978] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] sync_power_state_interval = 600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.062151] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] sync_power_state_pool_size = 1000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.062314] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] syslog_log_facility = LOG_USER {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.062468] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] tempdir = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.062620] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] thread_pool_statistic_period = -1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.062772] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] timeout_nbd = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.062933] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] transport_url = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.063109] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] update_resources_interval = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.063287] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] use_cow_images = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.063444] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] use_journal = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.063595] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] use_json = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.063744] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] use_rootwrap_daemon = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.063893] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] use_stderr = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064055] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] use_syslog = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064208] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vcpu_pin_set = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064372] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plugging_is_fatal = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064534] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plugging_timeout = 300 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064695] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] virt_mkfs = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064846] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] volume_usage_poll_interval = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.064998] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] watch_log_file = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.065171] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] web = /usr/share/spice-html5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 532.065417] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_concurrency.disable_process_locking = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.066063] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.066322] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.066503] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.066667] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.066832] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.066991] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.067176] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.067343] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.067503] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.067686] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.compute_link_prefix = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.067855] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068029] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.dhcp_domain = novalocal {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068193] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.enable_instance_password = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068351] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.glance_link_prefix = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068507] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068670] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068830] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.instance_list_per_project_cells = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.068983] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.list_records_by_skipping_down_cells = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.069156] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.local_metadata_per_cell = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.069315] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.max_limit = 1000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.069472] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.metadata_cache_expiration = 15 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.069639] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.neutron_default_project_id = default {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.069803] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.response_validation = warn {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.069959] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.use_neutron_default_nets = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.070130] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.070287] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.070439] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.070599] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.070760] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_dynamic_targets = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.070917] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_jsonfile_path = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.071099] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.071290] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.backend = dogpile.cache.memcached {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.071454] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.backend_argument = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.071607] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.backend_expiration_time = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.071767] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.config_prefix = cache.oslo {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.071922] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.debug_cache_backend = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.072086] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.enable_retry_client = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.072243] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.enable_socket_keepalive = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.072405] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.enabled = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.072558] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.enforce_fips_mode = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.072710] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.expiration_time = 600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.072878] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.hashclient_dead_timeout = 60.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.073038] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.hashclient_retry_attempts = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.073234] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.hashclient_retry_timeout = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.073398] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_dead_retry = 300 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.073550] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_password = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.073705] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.073858] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074025] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_pool_maxsize = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074178] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074335] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_sasl_enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074508] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074668] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_socket_timeout = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074819] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.memcache_username = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.074975] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.proxies = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.075143] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_db = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.075380] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_password = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.075569] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_sentinel_service_name = mymaster {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.075741] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.075902] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_server = localhost:6379 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.076075] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_socket_timeout = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.076232] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.redis_username = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.076389] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.retry_attempts = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.076547] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.retry_delay = 0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.076700] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.socket_keepalive_count = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.076853] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.socket_keepalive_idle = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077014] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.socket_keepalive_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077174] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.tls_allowed_ciphers = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077324] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.tls_cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077476] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.tls_certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077650] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.tls_enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077806] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cache.tls_keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.077971] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.078154] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.auth_type = password {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.078309] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.078481] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.catalog_info = volumev3::publicURL {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.078632] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.078785] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.078936] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.cross_az_attach = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.079102] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.debug = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.079261] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.endpoint_template = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.079419] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.http_retries = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.079573] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.079724] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.079891] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.os_region_name = RegionOne {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.080059] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.080224] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cinder.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.080396] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.080549] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.cpu_dedicated_set = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.080702] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.cpu_shared_set = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.080860] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.image_type_exclude_list = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.081115] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.081282] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.max_concurrent_disk_ops = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.081439] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.max_disk_devices_to_attach = -1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.081593] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.081752] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.081905] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.resource_provider_association_refresh = 300 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.082068] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.082226] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.shutdown_retry_interval = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.082398] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.082569] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] conductor.workers = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.082733] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] console.allowed_origins = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.082884] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] console.ssl_ciphers = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.083056] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] console.ssl_minimum_version = default {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.083255] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] consoleauth.enforce_session_timeout = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.083422] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] consoleauth.token_ttl = 600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.083588] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.083743] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.083902] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084065] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084223] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084375] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084529] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084677] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084829] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.084977] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.085140] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.085317] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.085477] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.085640] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.service_type = accelerator {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.085794] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.085984] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.086186] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.086348] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.086527] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.086680] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] cyborg.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.086844] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.asyncio_connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.086999] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.asyncio_slave_connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.087177] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.backend = sqlalchemy {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.087377] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.087578] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.connection_debug = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.087750] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.connection_parameters = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.087911] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.connection_recycle_time = 3600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.088079] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.connection_trace = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.088240] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.db_inc_retry_interval = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.088398] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.db_max_retries = 20 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.088554] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.db_max_retry_interval = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.088710] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.db_retry_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.088865] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.max_overflow = 50 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089028] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.max_pool_size = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089188] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.max_retries = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089356] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089508] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.mysql_wsrep_sync_wait = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089657] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.pool_timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089810] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.retry_interval = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.089962] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.slave_connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.090133] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.sqlite_synchronous = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.090289] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] database.use_db_reconnect = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.090452] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.asyncio_connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.090607] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.asyncio_slave_connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.090770] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.backend = sqlalchemy {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.090933] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.091104] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.connection_debug = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.091270] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.connection_parameters = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.091429] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.connection_recycle_time = 3600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.091584] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.connection_trace = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.091738] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.db_inc_retry_interval = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.091891] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.db_max_retries = 20 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.092063] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.db_max_retry_interval = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.092223] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.db_retry_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.092380] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.max_overflow = 50 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.092532] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.max_pool_size = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.092689] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.max_retries = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.092915] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.093102] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.093281] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.pool_timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.093442] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.retry_interval = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.093596] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.slave_connection = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.093753] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] api_database.sqlite_synchronous = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.093922] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] devices.enabled_mdev_types = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.094103] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.094272] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ephemeral_storage_encryption.default_format = luks {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.094430] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ephemeral_storage_encryption.enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.094588] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.094753] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.api_servers = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.094910] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.095080] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.095280] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.095444] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.095600] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.095755] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.debug = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.095917] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.default_trusted_certificate_ids = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096084] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.enable_certificate_validation = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096320] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.enable_rbd_download = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096392] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096546] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096696] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096844] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.096991] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.097161] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.num_retries = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.097324] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.rbd_ceph_conf = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.097479] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.rbd_connect_timeout = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.097708] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.rbd_pool = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.097889] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.rbd_user = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098059] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098218] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098370] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098532] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.service_type = image {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098685] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098834] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.098989] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.099158] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.099330] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.099487] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.verify_glance_signatures = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.099640] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] glance.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.099796] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] guestfs.debug = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.099952] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.100119] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.auth_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.100272] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.100424] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.100580] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.100733] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.100882] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.101043] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.101204] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.101356] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.101509] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.101689] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.101862] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102022] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102180] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102345] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.service_type = shared-file-system {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102503] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.share_apply_policy_timeout = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102658] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102810] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.102960] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.103147] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.103340] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.103500] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] manila.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.103663] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] mks.enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.104048] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.104243] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] image_cache.manager_interval = 2400 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.104409] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] image_cache.precache_concurrency = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.104575] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] image_cache.remove_unused_base_images = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.104736] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.104895] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.105074] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] image_cache.subdirectory_name = _base {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.105276] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.api_max_retries = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.105440] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.api_retry_interval = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.105593] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.105751] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.auth_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.105906] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.106074] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.106245] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.106399] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.conductor_group = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.106552] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.106705] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.106857] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107016] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107174] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107325] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107477] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107662] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.peer_list = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107818] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.107967] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.108140] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.serial_console_state_timeout = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.108294] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.108459] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.service_type = baremetal {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.108616] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.shard = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.108770] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.108921] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.109084] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.109240] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.109415] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.109569] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ironic.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.109747] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.109916] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] key_manager.fixed_key = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.110114] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.110274] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.barbican_api_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.110429] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.barbican_endpoint = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.110592] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.barbican_endpoint_type = public {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.110745] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.barbican_region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.110897] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111060] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111219] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111376] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111525] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111678] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.number_of_retries = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111830] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.retry_delay = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.111982] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.send_service_user_token = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.112150] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.112303] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.112458] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.verify_ssl = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.112611] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican.verify_ssl_path = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.112766] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.112918] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.auth_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.113078] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.113234] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.113390] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.113541] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.113713] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.113888] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114055] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] barbican_service_user.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114219] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.approle_role_id = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114372] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.approle_secret_id = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114533] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.kv_mountpoint = secret {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114685] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.kv_path = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114841] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.kv_version = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.114992] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.namespace = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.115158] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.root_token_id = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.115338] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.ssl_ca_crt_file = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.115505] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.timeout = 60.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.115660] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.use_ssl = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.115844] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116032] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116193] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116353] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116508] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116661] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116814] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.116966] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.117130] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.117282] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.117432] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.117606] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.117773] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.117924] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.118101] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.service_type = identity {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.118262] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.118418] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.118570] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.118721] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.118893] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.119055] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] keystone.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.119244] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.ceph_mount_options = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.119570] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.119742] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.connection_uri = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.119899] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_mode = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.120069] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_model_extra_flags = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.120234] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_models = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.120400] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_power_governor_high = performance {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.120562] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_power_governor_low = powersave {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.120721] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_power_management = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.120884] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.121049] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.device_detach_attempts = 8 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.121212] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.device_detach_timeout = 20 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.121372] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.disk_cachemodes = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.121526] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.disk_prefix = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.121683] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.enabled_perf_events = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.121850] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.file_backed_memory = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122025] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.gid_maps = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122189] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.hw_disk_discard = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122346] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.hw_machine_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122508] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_rbd_ceph_conf = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122664] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122818] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.122977] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_rbd_glance_store_name = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.123153] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_rbd_pool = rbd {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.123319] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_type = default {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.123473] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.images_volume_group = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.123628] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.inject_key = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.123782] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.inject_partition = -2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.123938] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.inject_password = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.124105] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.iscsi_iface = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.124263] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.iser_use_multipath = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.124418] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_bandwidth = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.124573] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.124725] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_downtime = 500 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.124877] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.125047] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.125287] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_inbound_addr = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.125465] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_parallel_connections = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.125628] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.125807] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_permit_post_copy = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.125964] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_scheme = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.126148] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_timeout_action = abort {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.126302] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_tunnelled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.126455] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_uri = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.126608] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.live_migration_with_native_tls = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.126757] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.max_queues = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.126911] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.127159] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.127319] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.nfs_mount_options = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.127660] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.127872] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.128055] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.num_iser_scan_tries = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.128218] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.num_memory_encrypted_guests = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.128377] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.128535] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.num_pcie_ports = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.128695] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.num_volume_scan_tries = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.128853] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.pmem_namespaces = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.129019] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.quobyte_client_cfg = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.129327] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.129501] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rbd_connect_timeout = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.129659] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.129815] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.129970] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rbd_secret_uuid = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.130135] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rbd_user = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.130292] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.130469] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.remote_filesystem_transport = ssh {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.130674] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rescue_image_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.130835] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rescue_kernel_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.130986] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rescue_ramdisk_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.131163] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.131314] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.rx_queue_size = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.131474] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.smbfs_mount_options = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.131767] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.131935] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.snapshot_compression = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.132104] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.snapshot_image_format = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.132334] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.132495] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.sparse_logical_volumes = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.132673] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.supported_tpm_secret_security = ['user', 'host'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.132830] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.swtpm_enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.132991] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.swtpm_group = tss {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.133166] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.swtpm_user = tss {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.133366] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.sysinfo_serial = unique {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.133530] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.tb_cache_size = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.133685] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.tx_queue_size = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.133845] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.uid_maps = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134019] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.use_default_aio_mode_for_volumes = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134172] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.use_virtio_for_bridges = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134337] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.virt_type = kvm {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134501] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.volume_clear = zero {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134657] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.volume_clear_size = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134811] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.volume_enforce_multipath = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.134967] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.volume_use_multipath = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.135132] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_cache_path = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.135328] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.135499] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_mount_group = qemu {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.135659] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_mount_opts = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.135820] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.136117] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.136292] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.vzstorage_mount_user = stack {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.136456] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.136623] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.136789] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.auth_type = password {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.136941] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.137104] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.137260] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.137410] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.137583] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.137759] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.default_floating_pool = public {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.137912] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.138089] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.extension_sync_interval = 600 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.138248] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.http_retries = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.138402] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.138554] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.138706] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.138866] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139029] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139201] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.ovs_bridge = br-int {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139362] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.physnets = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139522] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.region_name = RegionOne {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139673] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139833] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.service_metadata_proxy = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.139983] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.140157] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.service_type = network {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.140312] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.140463] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.140613] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.140761] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.140928] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.141090] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] neutron.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.141255] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] notifications.bdms_in_notifications = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.141425] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] notifications.default_level = INFO {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.141581] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] notifications.include_share_mapping = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.141747] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] notifications.notification_format = unversioned {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.141901] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] notifications.notify_on_state_change = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.142081] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.142254] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] pci.alias = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.142416] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] pci.device_spec = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.142572] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] pci.report_in_placement = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.142737] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.142899] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.auth_type = password {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143067] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143222] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143372] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143526] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143676] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143825] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.143974] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.default_domain_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.144136] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.default_domain_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.144287] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.domain_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.144435] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.domain_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.144582] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.144734] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.144881] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.145037] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.145224] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.145380] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.password = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.145535] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.project_domain_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.145696] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.project_domain_name = Default {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.145851] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.project_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146023] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.project_name = service {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146199] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.region_name = RegionOne {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146364] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146519] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146679] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.service_type = placement {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146835] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.146984] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.147150] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.147299] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.system_scope = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.147448] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.147629] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.trust_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.147789] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.user_domain_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.147951] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.user_domain_name = Default {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.148120] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.user_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.148288] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.username = nova {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.148459] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.148613] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] placement.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.148790] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.cores = 20 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.148947] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.count_usage_from_placement = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.149145] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.149313] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.injected_file_content_bytes = 10240 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.149474] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.injected_file_path_length = 255 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.149630] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.injected_files = 5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.149784] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.instances = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.149940] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.key_pairs = 100 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.150109] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.metadata_items = 128 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.150266] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.ram = 51200 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.150421] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.recheck_quota = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.150577] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.server_group_members = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.150731] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.server_groups = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.150929] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.151114] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] quota.unified_limits_resource_strategy = require {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.151282] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.151440] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.151592] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.image_metadata_prefilter = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.151742] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.151895] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.max_attempts = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.152061] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.max_placement_results = 1000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.152219] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.152373] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.query_placement_for_image_type_support = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.152536] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.152701] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] scheduler.workers = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.152868] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.153059] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.153240] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.153405] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.153562] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.153720] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.153875] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154063] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154223] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.host_subset_size = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154381] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154530] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154682] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154831] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.154981] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.155146] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.isolated_hosts = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.155323] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.isolated_images = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.155494] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.155645] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.155798] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.155950] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.pci_in_placement = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.156118] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.156294] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.156461] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.156623] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.156778] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.156928] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.157094] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.track_instance_changes = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.157261] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.157420] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] metrics.required = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.157596] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] metrics.weight_multiplier = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.157767] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.157926] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] metrics.weight_setting = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.158253] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.158421] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] serial_console.enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.158591] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] serial_console.port_range = 10000:20000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.158754] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.158921] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.159103] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] serial_console.serialproxy_port = 6083 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.159268] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.159431] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.auth_type = password {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.159597] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.159762] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.159917] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.160082] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.160235] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.160397] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.send_service_user_token = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.160551] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.160702] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] service_user.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.160861] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.agent_enabled = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.161024] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.161340] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.161541] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.161699] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.html5proxy_port = 6082 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.161850] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.image_compression = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.162012] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.jpeg_compression = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.162164] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.playback_compression = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.162317] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.require_secure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.162476] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.server_listen = 127.0.0.1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.162634] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.162968] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.163473] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.streaming_mode = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.163647] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] spice.zlib_compression = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.163812] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] upgrade_levels.baseapi = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.163978] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] upgrade_levels.compute = auto {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.164149] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] upgrade_levels.conductor = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.164301] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] upgrade_levels.scheduler = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.164459] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.164612] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.auth_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.164761] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.164912] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.165080] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.165269] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.165426] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.165581] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.165730] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vendordata_dynamic_auth.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.165894] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.api_retry_count = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.166056] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.ca_file = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.166227] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.cache_prefix = devstack-image-cache {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.166388] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.cluster_name = testcl1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.166544] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.connection_pool_size = 10 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.166697] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.console_delay_seconds = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.166858] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.datastore_regex = ^datastore.* {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.167088] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.167270] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.host_password = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.167432] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.host_port = 443 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.167616] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.host_username = administrator@vsphere.local {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.167786] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.insecure = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.167942] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.integration_bridge = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.168117] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.maximum_objects = 100 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.168272] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.pbm_default_policy = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.168429] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.pbm_enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.168578] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.pbm_wsdl_location = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.168740] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.168890] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.serial_port_proxy_uri = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.169051] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.serial_port_service_uri = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.169219] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.task_poll_interval = 0.5 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.169470] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.use_linked_clone = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.169595] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.vnc_keymap = en-us {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.169754] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.vnc_port = 5900 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.169910] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vmware.vnc_port_total = 10000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.170098] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.auth_schemes = ['none'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.170271] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.170552] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.170726] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.170891] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.novncproxy_port = 6080 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.171084] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.server_listen = 127.0.0.1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.171260] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.171417] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.vencrypt_ca_certs = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.171568] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.vencrypt_client_cert = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.171749] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vnc.vencrypt_client_key = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.171920] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172086] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.disable_deep_image_inspection = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172241] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172392] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172541] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172692] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.disable_rootwrap = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172840] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.enable_numa_live_migration = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.172990] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.173170] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.173309] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.173458] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.libvirt_disable_apic = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.173604] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.173752] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.173900] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174058] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174210] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174359] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174504] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174650] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174795] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.174945] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.175132] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.175316] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] wsgi.secure_proxy_ssl_header = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.175474] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] zvm.ca_file = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.175628] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] zvm.cloud_connector_url = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.175928] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.176118] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] zvm.reachable_timeout = 300 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.176287] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.176458] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.176627] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.connection_string = messaging:// {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.176788] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.enabled = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.176946] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.es_doc_type = notification {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.177113] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.es_scroll_size = 10000 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.177273] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.es_scroll_time = 2m {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.177429] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.filter_error_trace = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.177612] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.hmac_keys = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.177783] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.sentinel_service_name = mymaster {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.177936] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.socket_timeout = 0.1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.178108] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.trace_requests = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.178264] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler.trace_sqlalchemy = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.178434] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler_jaeger.process_tags = {} {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.178586] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler_jaeger.service_name_prefix = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.178738] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] profiler_otlp.service_name_prefix = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.178899] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179061] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179217] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179369] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179521] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179671] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179821] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.179968] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.180132] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.180293] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.180621] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.180621] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.180742] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.180894] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.181055] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.181215] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.181366] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.181514] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.181669] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.181818] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182054] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182228] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182380] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182591] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182673] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182821] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.182969] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.183133] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.183285] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.183439] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.183591] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.ssl = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.183778] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.183945] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.184116] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.184278] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.184437] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.ssl_version = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.184592] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.184776] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.184934] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_notifications.retry = -1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.185116] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.185318] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_messaging_notifications.transport_url = **** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.185489] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.auth_section = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.185647] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.auth_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.185799] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.cafile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.185951] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.certfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.186120] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.collect_timing = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.186294] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.connect_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.186450] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.connect_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.186601] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.endpoint_id = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.186762] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.endpoint_interface = publicURL {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.186911] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.endpoint_override = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.187079] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.endpoint_region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.187232] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.endpoint_service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.187382] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.endpoint_service_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.187549] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.insecure = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.187711] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.keyfile = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.187862] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.max_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188022] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.min_version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188179] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.region_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188332] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.retriable_status_codes = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188482] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.service_name = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188629] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.service_type = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188782] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.split_loggers = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.188932] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.status_code_retries = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.189095] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.status_code_retry_delay = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.189250] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.timeout = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.189400] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.valid_interfaces = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.189547] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_limit.version = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.189702] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_reports.file_event_handler = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.189858] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190013] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] oslo_reports.log_dir = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190189] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190343] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190492] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190644] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190800] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.190952] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.191113] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.191276] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.191484] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.group = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.191661] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.191819] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.191974] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.192143] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.192295] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] vif_plug_ovs_privileged.user = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.192459] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.flat_interface = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.192630] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.192793] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.192955] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.193137] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.193297] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.193455] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.193607] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.193775] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.193936] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.isolate_vif = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.194112] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.194270] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.194456] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.194625] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.ovsdb_interface = native {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.194779] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] os_vif_ovs.per_port_bridge = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.194941] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.capabilities = [21, 2] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.195109] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.group = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.195282] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.helper_command = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.195443] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.195597] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.195748] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.195895] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] privsep_osbrick.user = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196069] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196222] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.group = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196371] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.helper_command = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196521] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196673] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196824] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.196973] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] nova_sys_admin.user = None {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 532.197109] env[65107]: DEBUG oslo_service.backend._eventlet.service [None req-16ecbd61-4792-498c-bccb-f80d6fdece82 None None] ******************************************************************************** {{(pid=65107) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 532.197534] env[65107]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 532.702363] env[65107]: WARNING nova.virt.vmwareapi.driver [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 532.704145] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Getting list of instances from cluster (obj){ [ 532.704145] env[65107]: value = "domain-c8" [ 532.704145] env[65107]: _type = "ClusterComputeResource" [ 532.704145] env[65107]: } {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 532.704298] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1ca11d-dc29-4eaa-a2b8-96ef1d040912 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.713264] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Got total of 0 instances {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 532.713747] env[65107]: INFO nova.virt.node [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Generated node identity 8adccf8d-6d73-43c0-ad15-ee89967ab8cd [ 532.713979] env[65107]: INFO nova.virt.node [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Wrote node identity 8adccf8d-6d73-43c0-ad15-ee89967ab8cd to /opt/stack/data/n-cpu-1/compute_id [ 533.216770] env[65107]: WARNING nova.compute.manager [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Compute nodes ['8adccf8d-6d73-43c0-ad15-ee89967ab8cd'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 534.222331] env[65107]: INFO nova.compute.manager [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 535.228236] env[65107]: WARNING nova.compute.manager [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 535.228590] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 535.228659] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 535.228785] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 535.228920] env[65107]: DEBUG nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 535.229859] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1748e827-7524-4f0a-bfff-61d05633795b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.238351] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2958f244-43c0-463f-ba1d-27720748c1fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.253578] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be05da0-0382-40fb-b696-4e7fe34f3d95 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.260819] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20f240f-f9b0-49f2-993d-811437d1d802 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.291409] env[65107]: DEBUG nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180525MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 535.291597] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 535.291785] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 535.793920] env[65107]: WARNING nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] No compute node record for cpu-1:8adccf8d-6d73-43c0-ad15-ee89967ab8cd: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8adccf8d-6d73-43c0-ad15-ee89967ab8cd could not be found. [ 536.298306] env[65107]: INFO nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd [ 537.806482] env[65107]: DEBUG nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 537.806851] env[65107]: DEBUG nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 537.972650] env[65107]: INFO nova.scheduler.client.report [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] [req-2d008634-6554-4994-9b57-b15365e77593] Created resource provider record via placement API for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 537.989798] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e20173-b732-4696-adf4-88164ab6c388 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.999200] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f191f4-4647-4b52-8885-9da5358702c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.029573] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8f3837-7bc9-4d3a-b269-58cda4a6b6c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.037793] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a6ca99-8282-4778-b176-7533b59a9774 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.052748] env[65107]: DEBUG nova.compute.provider_tree [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 538.593054] env[65107]: DEBUG nova.scheduler.client.report [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 538.593054] env[65107]: DEBUG nova.compute.provider_tree [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 0 to 1 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 538.593054] env[65107]: DEBUG nova.compute.provider_tree [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 538.644147] env[65107]: DEBUG nova.compute.provider_tree [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 1 to 2 during operation: update_traits {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 539.148458] env[65107]: DEBUG nova.compute.resource_tracker [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 539.148458] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.854s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 539.148458] env[65107]: DEBUG nova.service [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Creating RPC server for service compute {{(pid=65107) start /opt/stack/nova/nova/service.py:177}} [ 539.161466] env[65107]: DEBUG nova.service [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] Join ServiceGroup membership for this service compute {{(pid=65107) start /opt/stack/nova/nova/service.py:194}} [ 539.161466] env[65107]: DEBUG nova.servicegroup.drivers.db [None req-43f027dd-0cf9-41e5-93f7-becff12af2d6 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=65107) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 555.165170] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_power_states {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 555.668164] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Getting list of instances from cluster (obj){ [ 555.668164] env[65107]: value = "domain-c8" [ 555.668164] env[65107]: _type = "ClusterComputeResource" [ 555.668164] env[65107]: } {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 555.669811] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ceb1d8-cb49-444f-9421-5b9fc26ea986 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.678861] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Got total of 0 instances {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 555.679117] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 555.679413] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Getting list of instances from cluster (obj){ [ 555.679413] env[65107]: value = "domain-c8" [ 555.679413] env[65107]: _type = "ClusterComputeResource" [ 555.679413] env[65107]: } {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 555.680285] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cca25f-17c2-4380-8aed-9d542a3aeea0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.687988] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Got total of 0 instances {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 586.479368] env[65107]: INFO nova.utils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] The default thread pool MainProcess.default is initialized [ 586.479949] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 586.480064] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 586.760794] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "e3510c31-d6be-4e9f-a0a2-a662123861e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 586.761302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 586.983018] env[65107]: DEBUG nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 587.264497] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 587.510335] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.510838] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.511427] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.512744] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.513211] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.513428] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.513714] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.514231] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 587.514231] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 587.551937] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 587.552387] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 587.554996] env[65107]: INFO nova.compute.claims [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.787040] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "31195587-86ce-4379-946d-1abb51c35f08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 587.787040] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "31195587-86ce-4379-946d-1abb51c35f08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 587.810868] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.021462] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.213960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.214307] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 588.292018] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 588.340999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "b779b5ad-7d93-47c2-b824-6d76246c00f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.342900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 588.699744] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2816a5d-0627-4843-b32b-5911e5c97f37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.708586] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cb4919-12ac-4f3c-8375-4663304a8478 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.750059] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 588.754281] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38661387-d8b5-4bae-8ae3-fbdf057a7a10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.762883] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57f23ed-051f-4d29-af07-cfe9e62fae59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.778428] env[65107]: DEBUG nova.compute.provider_tree [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.827221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.846235] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 589.112567] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.114704] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 589.283034] env[65107]: DEBUG nova.scheduler.client.report [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 589.288308] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.385412] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.616298] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 589.793526] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 589.793526] env[65107]: DEBUG nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 589.794861] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.987s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 589.797480] env[65107]: INFO nova.compute.claims [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.945306] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "0ef89aea-5373-4d07-9587-37890c08d35b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.945553] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "0ef89aea-5373-4d07-9587-37890c08d35b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 590.140415] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 590.308280] env[65107]: DEBUG nova.compute.utils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 590.309254] env[65107]: DEBUG nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 590.452016] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 590.811747] env[65107]: DEBUG nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 590.968023] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede0d234-0688-439b-b18d-00cd3b1d76c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.976150] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f80491e-ddfd-4020-a200-132add467f6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.981319] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.012735] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab51838-d341-44d3-a5a0-a08c9cb1d062 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.022178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b866a7-34c4-443a-a65d-759cac96d833 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.039566] env[65107]: DEBUG nova.compute.provider_tree [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.546042] env[65107]: DEBUG nova.scheduler.client.report [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 591.824685] env[65107]: DEBUG nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 592.050064] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 592.050600] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 592.053509] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.032s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.053509] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 592.053683] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 592.053913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.227s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.055439] env[65107]: INFO nova.compute.claims [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.061123] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891fa22a-f5c8-46c6-8057-15eeab796e74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.067485] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e1b016-5e98-4c50-b765-c297a993caca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.083427] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb993b85-6dbd-44fa-8842-ef2481c9376f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.091423] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815e2b43-d5c8-487e-874e-a42334e8bd9f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.121544] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180525MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 592.121692] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.168038] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 592.168584] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 592.168584] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 592.168700] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 592.168736] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 592.168892] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 592.169102] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.169255] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 592.169695] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 592.169859] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 592.170047] env[65107]: DEBUG nova.virt.hardware [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 592.171046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e156015a-46d5-4b7d-a0c8-7191ccf8401f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.179352] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b659af6-315c-43df-954f-d48159328f73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.200855] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e959149-f071-4f79-8194-c7979d6593bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.217880] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 592.229022] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 592.229022] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b24b1db-7bfa-4b48-8bba-c2e22dd8a365 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.242750] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Created folder: OpenStack in parent group-v4. [ 592.242980] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating folder: Project (1134e7eca4244c8b8b038a72efa4210f). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 592.243328] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2e6dd33-1c12-47a9-9d0d-3d99b3faae54 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.255271] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Created folder: Project (1134e7eca4244c8b8b038a72efa4210f) in parent group-v992574. [ 592.255271] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating folder: Instances. Parent ref: group-v992575. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 592.255589] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a54c95b9-195c-48f0-bcfd-e9df484b2f07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.265078] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Created folder: Instances in parent group-v992575. [ 592.266050] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 592.266050] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 592.266050] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99fba9de-897a-442d-9915-40801fe490b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.286893] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 592.286893] env[65107]: value = "task-5102140" [ 592.286893] env[65107]: _type = "Task" [ 592.286893] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.302628] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102140, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.560057] env[65107]: DEBUG nova.compute.utils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 592.561802] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 592.562120] env[65107]: DEBUG nova.network.neutron [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 592.563098] env[65107]: WARNING neutronclient.v2_0.client [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.565611] env[65107]: WARNING neutronclient.v2_0.client [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.567388] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 592.568147] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 592.798220] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102140, 'name': CreateVM_Task, 'duration_secs': 0.337324} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.798478] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 592.799577] env[65107]: DEBUG oslo_vmware.service [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fd2c37-e16d-4c4d-95ff-ad09af95e58e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.805960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.806136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 592.806789] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 592.807065] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53efbaa4-d2ca-4bbb-818a-d4e42ce834a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.812050] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 592.812050] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52607924-562e-9f99-07fa-8b1c1d89cca3" [ 592.812050] env[65107]: _type = "Task" [ 592.812050] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.820810] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52607924-562e-9f99-07fa-8b1c1d89cca3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.077481] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 593.187020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed6d8e7-5082-4a96-a545-5eac57d50cb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.194167] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8c0b7a-067a-4d1d-b2bd-fe1b84f31526 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.232589] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fa7cb2-6d96-4db5-bc14-d6d556469e65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.241353] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c8bbf3-d985-49ba-a43e-7987ad0f99ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.255728] env[65107]: DEBUG nova.compute.provider_tree [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.323898] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 593.323898] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 593.323898] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.324267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 593.324823] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 593.325208] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-956f8128-23ad-4cc2-a95d-f94e395dc23b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.337056] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 593.337056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 593.337056] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c17cb74-529a-4d9b-b3ca-8008c4680246 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.347242] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-165e093b-1961-46e0-bd30-cf385bc01248 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.352465] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 593.352465] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523b30f9-2a61-9f63-8c57-9850cc0e1b81" [ 593.352465] env[65107]: _type = "Task" [ 593.352465] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.361346] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523b30f9-2a61-9f63-8c57-9850cc0e1b81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.482578] env[65107]: DEBUG nova.policy [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd7f9abea8d946ecb2accee2714703e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2d4844feda0432ebe1ef160b0070b83', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 593.759606] env[65107]: DEBUG nova.scheduler.client.report [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 593.867232] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 593.867342] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating directory with path [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 593.867591] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abbc23c8-1d1f-4577-ab9a-79fdc272c684 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.890036] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Created directory with path [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 593.890338] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Fetch image to [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 593.890541] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Downloading image file data e439aeac-2298-462c-be63-8218195135cf to [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk on the data store datastore1 {{(pid=65107) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 593.892015] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c74dae-db4d-47bd-b316-e20c3e474570 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.900152] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3932cfb0-374a-4bdd-a345-fcf47b2287df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.910566] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097eff9f-e008-4381-9308-836b42dd7a9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.942913] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96dcb446-cb6e-45a2-822f-f051304eb136 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.949618] env[65107]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-5297efc8-22c4-496c-a0bc-daf3933b3abb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.985351] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Downloading image file data e439aeac-2298-462c-be63-8218195135cf to the data store datastore1 {{(pid=65107) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 594.070635] env[65107]: DEBUG oslo_vmware.rw_handles [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 594.135676] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 594.168857] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 594.169009] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 594.169240] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 594.169435] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 594.169619] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 594.169786] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 594.170116] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.170248] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 594.170442] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 594.170636] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 594.170839] env[65107]: DEBUG nova.virt.hardware [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 594.171971] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3894982e-f63c-46e3-b49d-f289e80784c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.183230] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb1eff4-8e9f-4027-9f8f-9e7a5e0695f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.265526] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.211s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 594.266065] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 594.269313] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.981s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 594.270818] env[65107]: INFO nova.compute.claims [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.759608] env[65107]: DEBUG oslo_vmware.rw_handles [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 594.759818] env[65107]: DEBUG oslo_vmware.rw_handles [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 594.776862] env[65107]: DEBUG nova.compute.utils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 594.778674] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 594.779706] env[65107]: DEBUG nova.network.neutron [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 594.779706] env[65107]: WARNING neutronclient.v2_0.client [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 594.779706] env[65107]: WARNING neutronclient.v2_0.client [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 594.783584] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.783584] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.791205] env[65107]: DEBUG nova.network.neutron [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Successfully created port: 2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 594.902796] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Downloaded image file data e439aeac-2298-462c-be63-8218195135cf to vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk on the data store datastore1 {{(pid=65107) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 594.905155] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 594.905155] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Copying Virtual Disk [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk to [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 594.905630] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-646843ed-c3dd-4452-9c89-0a841ef01a63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.917992] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 594.917992] env[65107]: value = "task-5102141" [ 594.917992] env[65107]: _type = "Task" [ 594.917992] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.931277] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.197855] env[65107]: DEBUG nova.policy [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '998c06c46eda4a84aa7ddd605721aaf3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aaf15dff8ae0472d94c9fecd0ba87a20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 595.297991] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 595.430993] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102141, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.432923] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e100a1-86ad-4f27-b42f-7a76d3376846 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.442347] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a404cc6-399a-44f2-a1b2-375828e50ec9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.481459] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11220816-baab-45af-8fc3-ba045257e942 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.491407] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f6c771-21d6-49c3-b382-a725733dfeaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.505779] env[65107]: DEBUG nova.compute.provider_tree [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.929931] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690837} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.930509] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Copied Virtual Disk [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk to [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 595.930711] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleting the datastore file [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 595.930978] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79fcdcff-9fb2-40ef-967c-e34a2f849677 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.939619] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 595.939619] env[65107]: value = "task-5102142" [ 595.939619] env[65107]: _type = "Task" [ 595.939619] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.950643] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102142, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.009620] env[65107]: DEBUG nova.scheduler.client.report [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 596.313777] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 596.345350] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 596.345776] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 596.346563] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 596.346877] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 596.347205] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 596.347377] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 596.347592] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.347747] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 596.347907] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 596.348073] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 596.348245] env[65107]: DEBUG nova.virt.hardware [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 596.349854] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af186e95-adec-4e54-b4e6-a097fe06e189 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.361203] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485821cb-33f2-410e-b33d-84c36dcfffc6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.452524] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102142, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026277} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.453258] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 596.453258] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Moving file from [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd/e439aeac-2298-462c-be63-8218195135cf to [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf. {{(pid=65107) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 596.453396] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-16a6911b-1edb-45be-96c8-99d6c0c6691a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.463576] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 596.463576] env[65107]: value = "task-5102143" [ 596.463576] env[65107]: _type = "Task" [ 596.463576] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.474698] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102143, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.514780] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.246s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 596.515402] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 596.522021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.133s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 596.522021] env[65107]: INFO nova.compute.claims [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.858454] env[65107]: DEBUG nova.network.neutron [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Successfully created port: 7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 596.907031] env[65107]: DEBUG nova.network.neutron [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Successfully updated port: 2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 596.976388] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102143, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.033526} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.976692] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] File moved {{(pid=65107) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 596.976887] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Cleaning up location [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 596.977061] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleting the datastore file [datastore1] vmware_temp/5355e016-33f9-45df-99af-6b36caa80dcd {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 596.977386] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9830290-172f-46ef-bb51-3c77d1e3c002 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.985538] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 596.985538] env[65107]: value = "task-5102144" [ 596.985538] env[65107]: _type = "Task" [ 596.985538] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.994359] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.025019] env[65107]: DEBUG nova.compute.utils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 597.035286] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 597.035286] env[65107]: DEBUG nova.network.neutron [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 597.035286] env[65107]: WARNING neutronclient.v2_0.client [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.035286] env[65107]: WARNING neutronclient.v2_0.client [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.035286] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.035526] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 597.162841] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5131aa8-8619-4c57-ae08-120c0e7304af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.171809] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0069f560-321b-4edc-a1cf-0b31b45bb633 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.205299] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc73598-7372-43a6-85a9-9bcbda048a78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.210850] env[65107]: DEBUG nova.policy [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a644dfd5e9614da292648b0cb98218c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4df06b84d1cd45b28b8ea33565ac45f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 597.219356] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca7429a-34d1-43ba-9a6e-6e9efb0c1f81 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.237451] env[65107]: DEBUG nova.compute.provider_tree [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.409487] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "refresh_cache-e3510c31-d6be-4e9f-a0a2-a662123861e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.409791] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "refresh_cache-e3510c31-d6be-4e9f-a0a2-a662123861e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 597.409831] env[65107]: DEBUG nova.network.neutron [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 597.497624] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026021} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.501103] env[65107]: DEBUG nova.network.neutron [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Successfully created port: d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 597.502483] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 597.504123] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2f5dceb-80c4-4165-9ef3-c46680d027b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.511962] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 597.511962] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d53d64-31ab-303e-7b48-dcea54a62eed" [ 597.511962] env[65107]: _type = "Task" [ 597.511962] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.522521] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d53d64-31ab-303e-7b48-dcea54a62eed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.534036] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 597.738089] env[65107]: DEBUG nova.scheduler.client.report [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.913056] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.913635] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.023923] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d53d64-31ab-303e-7b48-dcea54a62eed, 'name': SearchDatastore_Task, 'duration_secs': 0.012456} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.024190] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 598.024448] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 598.024697] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-734ee611-945e-4c7d-a4cc-8e5711c6a554 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.033598] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 598.033598] env[65107]: value = "task-5102145" [ 598.033598] env[65107]: _type = "Task" [ 598.033598] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.046904] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.246480] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.725s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 598.246480] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 598.247700] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.107s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 598.249243] env[65107]: INFO nova.compute.claims [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.547913] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466404} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.547913] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 598.547913] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 598.547913] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6238eea-6516-4f90-893b-619a93c21163 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.549037] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 598.557587] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 598.557587] env[65107]: value = "task-5102146" [ 598.557587] env[65107]: _type = "Task" [ 598.557587] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.564716] env[65107]: DEBUG nova.network.neutron [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Successfully updated port: 7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 598.577983] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102146, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.580569] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 598.581076] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 598.581076] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 598.581976] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 598.581976] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 598.581976] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 598.581976] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.582202] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 598.582444] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 598.582650] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 598.583071] env[65107]: DEBUG nova.virt.hardware [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 598.584099] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7ea032-f7eb-4eab-b74f-e795c4b3bc0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.594830] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d979588-d8ae-4f26-8d95-b8d3fee05d42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.750955] env[65107]: DEBUG nova.network.neutron [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 598.754283] env[65107]: DEBUG nova.compute.utils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 598.758254] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 598.758535] env[65107]: DEBUG nova.network.neutron [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 598.759671] env[65107]: WARNING neutronclient.v2_0.client [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 598.759671] env[65107]: WARNING neutronclient.v2_0.client [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 598.759921] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.760270] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.854442] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.855144] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.893681] env[65107]: DEBUG nova.policy [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd7f9abea8d946ecb2accee2714703e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2d4844feda0432ebe1ef160b0070b83', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 599.068596] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "refresh_cache-31195587-86ce-4379-946d-1abb51c35f08" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.068773] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired lock "refresh_cache-31195587-86ce-4379-946d-1abb51c35f08" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 599.068954] env[65107]: DEBUG nova.network.neutron [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 599.070342] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102146, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072601} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.070589] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 599.071768] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91dcb73c-3161-4259-b98e-8d41b3a9e9e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.094612] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 599.095551] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8b09a17-9092-4fc0-b988-1543a0ed4424 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.116936] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 599.116936] env[65107]: value = "task-5102147" [ 599.116936] env[65107]: _type = "Task" [ 599.116936] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.126383] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102147, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.243204] env[65107]: DEBUG nova.network.neutron [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Successfully created port: 3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 599.259384] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 599.278200] env[65107]: DEBUG nova.network.neutron [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Successfully updated port: d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 599.364366] env[65107]: WARNING neutronclient.v2_0.client [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.365051] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.365449] env[65107]: WARNING openstack [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.425935] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ddb67d-5ad3-4487-8dd0-e5dc5d802275 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.434678] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dab25bf-d482-4b4f-bec3-70ffabfc3b5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.469270] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1ebecf-a061-4d88-a398-abb5f5e33775 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.477188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1da412-5281-4e58-ba69-1c16fd0730af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.491443] env[65107]: DEBUG nova.compute.provider_tree [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.497712] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.497937] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "8306ff90-b8bd-4270-8133-96abe483156b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.572634] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.573163] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.627362] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102147, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.710183] env[65107]: DEBUG nova.network.neutron [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 599.783061] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.783302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquired lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 599.783485] env[65107]: DEBUG nova.network.neutron [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 599.974252] env[65107]: DEBUG nova.network.neutron [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Updating instance_info_cache with network_info: [{"id": "2686e3e0-c528-42bb-9444-050209faa6b1", "address": "fa:16:3e:89:7b:5d", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2686e3e0-c5", "ovs_interfaceid": "2686e3e0-c528-42bb-9444-050209faa6b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 599.994171] env[65107]: DEBUG nova.scheduler.client.report [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 600.000968] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 600.010966] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.011593] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 600.130088] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102147, 'name': ReconfigVM_Task, 'duration_secs': 0.660155} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.130343] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Reconfigured VM instance instance-00000001 to attach disk [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 600.131090] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b311bf45-dd19-4671-b68a-187f5846ac61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.137287] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 600.137287] env[65107]: value = "task-5102148" [ 600.137287] env[65107]: _type = "Task" [ 600.137287] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.146766] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102148, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.269822] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 600.286932] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.287633] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 600.305876] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 600.306140] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 600.306296] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 600.306479] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 600.306626] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 600.306768] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 600.306983] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.307155] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 600.307320] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 600.307510] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 600.307657] env[65107]: DEBUG nova.virt.hardware [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 600.308915] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db128485-ad7c-4fa4-9f8a-087f6422453b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.318363] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260c3cfc-8d0f-40b9-9555-752bdd13d369 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.477158] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "refresh_cache-e3510c31-d6be-4e9f-a0a2-a662123861e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 600.477551] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance network_info: |[{"id": "2686e3e0-c528-42bb-9444-050209faa6b1", "address": "fa:16:3e:89:7b:5d", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2686e3e0-c5", "ovs_interfaceid": "2686e3e0-c528-42bb-9444-050209faa6b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 600.478081] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:7b:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2686e3e0-c528-42bb-9444-050209faa6b1', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 600.486311] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating folder: Project (a2d4844feda0432ebe1ef160b0070b83). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.486593] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b6a8e4f-4a1d-4f2a-b5f1-dd34f3613be4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.498552] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 600.499078] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 600.502637] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created folder: Project (a2d4844feda0432ebe1ef160b0070b83) in parent group-v992574. [ 600.502828] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating folder: Instances. Parent ref: group-v992578. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.503401] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.522s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 600.504877] env[65107]: INFO nova.compute.claims [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.511771] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53596f09-1074-42a2-b767-e12bd90b139a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.528235] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created folder: Instances in parent group-v992578. [ 600.528572] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 600.528927] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 600.529016] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b44bb49-8935-4e71-a466-9f619724fb23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.545356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.550488] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 600.550488] env[65107]: value = "task-5102151" [ 600.550488] env[65107]: _type = "Task" [ 600.550488] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.559773] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102151, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.647481] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102148, 'name': Rename_Task, 'duration_secs': 0.137056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.647775] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 600.647987] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1969cd12-675a-4e47-b7fd-83359c3056fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.656885] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 600.656885] env[65107]: value = "task-5102152" [ 600.656885] env[65107]: _type = "Task" [ 600.656885] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.664649] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.758206] env[65107]: DEBUG nova.network.neutron [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 600.884890] env[65107]: DEBUG nova.network.neutron [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Successfully updated port: 3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 600.921170] env[65107]: WARNING neutronclient.v2_0.client [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 600.921170] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.921170] env[65107]: WARNING openstack [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.004902] env[65107]: DEBUG nova.compute.utils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 601.015019] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 601.015019] env[65107]: DEBUG nova.network.neutron [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 601.015019] env[65107]: WARNING neutronclient.v2_0.client [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.015019] env[65107]: WARNING neutronclient.v2_0.client [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.015019] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.015372] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.060352] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102151, 'name': CreateVM_Task, 'duration_secs': 0.392666} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.060531] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 601.061034] env[65107]: WARNING neutronclient.v2_0.client [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.168401] env[65107]: DEBUG oslo_vmware.api [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102152, 'name': PowerOnVM_Task, 'duration_secs': 0.467106} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.168656] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 601.169070] env[65107]: INFO nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Took 9.34 seconds to spawn the instance on the hypervisor. [ 601.169343] env[65107]: DEBUG nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 601.170203] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170c6fd5-634d-4a11-a7ac-3b18a838085d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.260492] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.260682] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.328020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.328020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.328020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 601.328020] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f943d944-e632-4f73-8c07-417d6d64fce8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.333598] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 601.333598] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523495d1-02da-7310-0b65-b87a03ec1ed0" [ 601.333598] env[65107]: _type = "Task" [ 601.333598] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.343819] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523495d1-02da-7310-0b65-b87a03ec1ed0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.390203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "refresh_cache-b779b5ad-7d93-47c2-b824-6d76246c00f5" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.390203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "refresh_cache-b779b5ad-7d93-47c2-b824-6d76246c00f5" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.390203] env[65107]: DEBUG nova.network.neutron [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 601.418597] env[65107]: DEBUG nova.network.neutron [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Updating instance_info_cache with network_info: [{"id": "7518822c-68c7-4bd3-8296-e61bd76671e2", "address": "fa:16:3e:60:c5:8a", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.217", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7518822c-68", "ovs_interfaceid": "7518822c-68c7-4bd3-8296-e61bd76671e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 601.451162] env[65107]: DEBUG nova.policy [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66f11f09dd98447bb9851bee92183dcc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2882c2c7949045519a146655bb694a3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 601.514097] env[65107]: WARNING neutronclient.v2_0.client [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.514767] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.515124] env[65107]: WARNING openstack [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.523052] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 601.692951] env[65107]: INFO nova.compute.manager [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Took 14.20 seconds to build instance. [ 601.707592] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be038ef-62ca-4963-990f-03a203b96053 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.716643] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5c2970-091d-47a6-beac-4e4607da38da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.769036] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0210061f-e2a0-4892-a5de-33195f404f3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.782343] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e2c1f1-e2a1-460d-a2eb-30729759090d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.798604] env[65107]: DEBUG nova.compute.provider_tree [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.850728] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523495d1-02da-7310-0b65-b87a03ec1ed0, 'name': SearchDatastore_Task, 'duration_secs': 0.031645} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.852978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.852978] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 601.852978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.852978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.853215] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 601.853215] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-734b9c3c-0603-4c91-b539-c0a000062189 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.866337] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 601.866426] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 601.867705] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c629fbb-1231-43d2-b733-06e55361099d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.876753] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 601.876753] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b79d79-3fee-46e5-eafa-63b1ddaf7aaa" [ 601.876753] env[65107]: _type = "Task" [ 601.876753] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.888855] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b79d79-3fee-46e5-eafa-63b1ddaf7aaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.891715] env[65107]: DEBUG nova.network.neutron [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Updating instance_info_cache with network_info: [{"id": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "address": "fa:16:3e:ae:a6:0b", "network": {"id": "4f3c2189-942b-47c1-95bd-5064d80e3a76", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-78304541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4df06b84d1cd45b28b8ea33565ac45f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0d2101e-2d93-4310-a242-af2d9ecdaf9b", "external-id": "nsx-vlan-transportzone-121", "segmentation_id": 121, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2435e56-5f", "ovs_interfaceid": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 601.895339] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.895339] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.921354] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Releasing lock "refresh_cache-31195587-86ce-4379-946d-1abb51c35f08" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.921700] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Instance network_info: |[{"id": "7518822c-68c7-4bd3-8296-e61bd76671e2", "address": "fa:16:3e:60:c5:8a", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.217", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7518822c-68", "ovs_interfaceid": "7518822c-68c7-4bd3-8296-e61bd76671e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 601.922161] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:c5:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7518822c-68c7-4bd3-8296-e61bd76671e2', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 601.934351] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Creating folder: Project (aaf15dff8ae0472d94c9fecd0ba87a20). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 601.935024] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e56315ab-bd0c-4338-a8de-6869a8776f92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.944530] env[65107]: DEBUG nova.network.neutron [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 601.950171] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Created folder: Project (aaf15dff8ae0472d94c9fecd0ba87a20) in parent group-v992574. [ 601.950364] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Creating folder: Instances. Parent ref: group-v992581. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 601.950644] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3038bae8-66e8-47c7-add4-f10a86b9e544 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.961726] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Created folder: Instances in parent group-v992581. [ 601.961726] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 601.961726] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 601.961914] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46e358a4-1e3f-463e-8ce5-fe7c4d455b80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.981745] env[65107]: DEBUG nova.network.neutron [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Successfully created port: 086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 601.986053] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 601.986053] env[65107]: value = "task-5102155" [ 601.986053] env[65107]: _type = "Task" [ 601.986053] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.994597] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102155, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.998233] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.998600] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.196183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-529b07f1-7a13-40f8-b981-7aaf7337ebe8 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.716s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 602.303844] env[65107]: DEBUG nova.scheduler.client.report [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 602.368853] env[65107]: WARNING neutronclient.v2_0.client [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.369523] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.369886] env[65107]: WARNING openstack [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.388702] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b79d79-3fee-46e5-eafa-63b1ddaf7aaa, 'name': SearchDatastore_Task, 'duration_secs': 0.012245} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.389427] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-296b2044-a833-4fae-9316-c3880009dbe2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.397704] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 602.397704] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5282ce3c-c825-ac4f-2de0-8cef533ecd92" [ 602.397704] env[65107]: _type = "Task" [ 602.397704] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.402518] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Releasing lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 602.403547] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Instance network_info: |[{"id": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "address": "fa:16:3e:ae:a6:0b", "network": {"id": "4f3c2189-942b-47c1-95bd-5064d80e3a76", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-78304541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4df06b84d1cd45b28b8ea33565ac45f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0d2101e-2d93-4310-a242-af2d9ecdaf9b", "external-id": "nsx-vlan-transportzone-121", "segmentation_id": 121, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2435e56-5f", "ovs_interfaceid": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 602.408681] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:a6:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a0d2101e-2d93-4310-a242-af2d9ecdaf9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2435e56-5f0a-44e9-8154-2f2dd4103c55', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 602.417343] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Creating folder: Project (4df06b84d1cd45b28b8ea33565ac45f1). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.417908] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5282ce3c-c825-ac4f-2de0-8cef533ecd92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.418581] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-54cc0e32-8080-42e5-9f87-0315cc946c4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.433930] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Created folder: Project (4df06b84d1cd45b28b8ea33565ac45f1) in parent group-v992574. [ 602.433930] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Creating folder: Instances. Parent ref: group-v992584. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.433930] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02f324c4-e43a-41ab-869a-37e6e3e6293a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.448592] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Created folder: Instances in parent group-v992584. [ 602.448848] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 602.449460] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 602.449460] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cde04d67-94fb-4df8-84e2-8bd0d5eac294 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.474963] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 602.474963] env[65107]: value = "task-5102158" [ 602.474963] env[65107]: _type = "Task" [ 602.474963] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.488263] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102158, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.499096] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102155, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.538307] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 602.567193] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 602.567749] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 602.567749] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 602.567938] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 602.568096] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 602.568256] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 602.568461] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.568630] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 602.568793] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 602.568957] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 602.569150] env[65107]: DEBUG nova.virt.hardware [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 602.570158] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0aac46-76ce-4e43-982d-ec32e7a93854 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.579954] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48daab6-7591-4661-8618-ed33cf59a583 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.690538] env[65107]: DEBUG nova.network.neutron [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Updating instance_info_cache with network_info: [{"id": "3ce4a5c1-5dae-42ba-919f-736e66893243", "address": "fa:16:3e:c0:0c:58", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ce4a5c1-5d", "ovs_interfaceid": "3ce4a5c1-5dae-42ba-919f-736e66893243", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 602.812415] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 602.812415] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 602.814714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.693s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 602.909679] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5282ce3c-c825-ac4f-2de0-8cef533ecd92, 'name': SearchDatastore_Task, 'duration_secs': 0.011477} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.909984] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 602.910757] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 602.910757] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ad1d22d-2db4-4ce9-8335-55b475a32bcc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.920966] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 602.920966] env[65107]: value = "task-5102159" [ 602.920966] env[65107]: _type = "Task" [ 602.920966] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.930643] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102159, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.987685] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102158, 'name': CreateVM_Task, 'duration_secs': 0.41009} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.987685] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 602.987685] env[65107]: WARNING neutronclient.v2_0.client [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.988812] env[65107]: DEBUG oslo_vmware.service [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7c4259-7af4-4d56-9f2c-0f75e4b5a008 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.001188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.001521] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 603.001865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 603.002488] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-354eae7b-e56f-42fa-b3e3-4347f853d6f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.011758] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 603.011758] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fc99f-1f33-2635-1f8f-5820ecd35ab1" [ 603.011758] env[65107]: _type = "Task" [ 603.011758] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.015692] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102155, 'name': CreateVM_Task, 'duration_secs': 0.571089} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.019112] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 603.019724] env[65107]: WARNING neutronclient.v2_0.client [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.019956] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.028710] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fc99f-1f33-2635-1f8f-5820ecd35ab1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.196752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "refresh_cache-b779b5ad-7d93-47c2-b824-6d76246c00f5" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 603.197031] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Instance network_info: |[{"id": "3ce4a5c1-5dae-42ba-919f-736e66893243", "address": "fa:16:3e:c0:0c:58", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ce4a5c1-5d", "ovs_interfaceid": "3ce4a5c1-5dae-42ba-919f-736e66893243", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 603.197536] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:0c:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ce4a5c1-5dae-42ba-919f-736e66893243', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 603.207565] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 603.207860] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 603.207999] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60bfa820-352a-47af-b840-5739aa1c455c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.231892] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 603.231892] env[65107]: value = "task-5102160" [ 603.231892] env[65107]: _type = "Task" [ 603.231892] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.242440] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102160, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.273892] env[65107]: DEBUG nova.compute.manager [req-d87fb130-1c16-4a99-991d-74e8031cec0c req-1d40e244-0dc5-4d63-bb31-e681da1be715 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Received event network-vif-plugged-2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 603.274379] env[65107]: DEBUG oslo_concurrency.lockutils [req-d87fb130-1c16-4a99-991d-74e8031cec0c req-1d40e244-0dc5-4d63-bb31-e681da1be715 service nova] Acquiring lock "e3510c31-d6be-4e9f-a0a2-a662123861e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 603.274646] env[65107]: DEBUG oslo_concurrency.lockutils [req-d87fb130-1c16-4a99-991d-74e8031cec0c req-1d40e244-0dc5-4d63-bb31-e681da1be715 service nova] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 603.274829] env[65107]: DEBUG oslo_concurrency.lockutils [req-d87fb130-1c16-4a99-991d-74e8031cec0c req-1d40e244-0dc5-4d63-bb31-e681da1be715 service nova] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 603.275100] env[65107]: DEBUG nova.compute.manager [req-d87fb130-1c16-4a99-991d-74e8031cec0c req-1d40e244-0dc5-4d63-bb31-e681da1be715 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] No waiting events found dispatching network-vif-plugged-2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 603.275359] env[65107]: WARNING nova.compute.manager [req-d87fb130-1c16-4a99-991d-74e8031cec0c req-1d40e244-0dc5-4d63-bb31-e681da1be715 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Received unexpected event network-vif-plugged-2686e3e0-c528-42bb-9444-050209faa6b1 for instance with vm_state building and task_state spawning. [ 603.322159] env[65107]: DEBUG nova.compute.utils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 603.328496] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 603.328716] env[65107]: DEBUG nova.network.neutron [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 603.329072] env[65107]: WARNING neutronclient.v2_0.client [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.329392] env[65107]: WARNING neutronclient.v2_0.client [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.329977] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.330806] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.437353] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102159, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.530285] env[65107]: DEBUG nova.policy [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba12978880904fd2a8d63adecec18409', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ae57a7184354c6b97c25d47135bcd40', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 603.541833] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 603.542140] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 603.542420] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.542583] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 603.542760] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 603.543402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 603.543888] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 603.544194] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63e82914-6dfd-459a-bc24-05a3eae373e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.546600] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e1f1d1-e738-4ec8-ac67-90e39777b460 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.553101] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 603.553101] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b598ec-b6ea-834b-de5c-9799f2d5711b" [ 603.553101] env[65107]: _type = "Task" [ 603.553101] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.557469] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 603.557649] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 603.559938] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dafa53d0-4cf9-4472-bc5f-9f25325a1efa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.566219] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b598ec-b6ea-834b-de5c-9799f2d5711b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.571390] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa337149-dfda-4514-8ba3-36104e124087 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.578113] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 603.578113] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52891547-5268-84f9-051d-0727801b0fb8" [ 603.578113] env[65107]: _type = "Task" [ 603.578113] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.589337] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52891547-5268-84f9-051d-0727801b0fb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.663961] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "ccd68268-d0fc-406e-896f-c61b114dd75d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 603.664282] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 603.683997] env[65107]: DEBUG nova.network.neutron [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Successfully updated port: 086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 603.746869] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102160, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.829085] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 603.868011] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 603.868208] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 0ef89aea-5373-4d07-9587-37890c08d35b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 603.935167] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102159, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647602} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.935475] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 603.935688] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 603.935941] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e3ea9f1-e9fa-4ab0-a87e-71fa005eb1dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.944194] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 603.944194] env[65107]: value = "task-5102161" [ 603.944194] env[65107]: _type = "Task" [ 603.944194] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.961242] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102161, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.991110] env[65107]: DEBUG nova.network.neutron [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Successfully created port: 74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 604.000593] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.000975] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 604.035415] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.035792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 604.065443] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 604.065830] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 604.066075] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.088945] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 604.089125] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Creating directory with path [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 604.089299] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2671b3dc-2ae8-4e0f-8bb6-cc9e49f858c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.115989] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Created directory with path [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 604.116563] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Fetch image to [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 604.116644] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Downloading image file data e439aeac-2298-462c-be63-8218195135cf to [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk on the data store datastore2 {{(pid=65107) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 604.117496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c5a07d-a6f7-4413-9300-29584cb62eaa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.126133] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbc2eb5-2e43-48b6-8a0b-8ff3f8b7734a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.137552] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fa7487-e841-44a7-be24-bcfa537dded3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.172200] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 604.178408] env[65107]: DEBUG nova.compute.manager [req-4878539e-a521-4f99-9065-91f2e8c9d4e9 req-8fa0836c-480f-4127-85ed-e4d5510a26af service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Received event network-vif-plugged-d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 604.178408] env[65107]: DEBUG oslo_concurrency.lockutils [req-4878539e-a521-4f99-9065-91f2e8c9d4e9 req-8fa0836c-480f-4127-85ed-e4d5510a26af service nova] Acquiring lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.178408] env[65107]: DEBUG oslo_concurrency.lockutils [req-4878539e-a521-4f99-9065-91f2e8c9d4e9 req-8fa0836c-480f-4127-85ed-e4d5510a26af service nova] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 604.178408] env[65107]: DEBUG oslo_concurrency.lockutils [req-4878539e-a521-4f99-9065-91f2e8c9d4e9 req-8fa0836c-480f-4127-85ed-e4d5510a26af service nova] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 604.178408] env[65107]: DEBUG nova.compute.manager [req-4878539e-a521-4f99-9065-91f2e8c9d4e9 req-8fa0836c-480f-4127-85ed-e4d5510a26af service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] No waiting events found dispatching network-vif-plugged-d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 604.178640] env[65107]: WARNING nova.compute.manager [req-4878539e-a521-4f99-9065-91f2e8c9d4e9 req-8fa0836c-480f-4127-85ed-e4d5510a26af service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Received unexpected event network-vif-plugged-d2435e56-5f0a-44e9-8154-2f2dd4103c55 for instance with vm_state building and task_state spawning. [ 604.178869] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9186776f-49c4-4997-9158-6ee90d49886c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.186924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.187095] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.187254] env[65107]: DEBUG nova.network.neutron [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 604.188277] env[65107]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-82d38f6b-7568-4437-a0c8-9b0ef06da0c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.224689] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Downloading image file data e439aeac-2298-462c-be63-8218195135cf to the data store datastore2 {{(pid=65107) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 604.246988] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102160, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.300139] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 604.372217] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance abb9dc0e-ce18-49c8-a472-d39a3fcec887 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 604.457277] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102161, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.189916} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.457570] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 604.458460] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d9daa7-c623-41a5-a20b-ceb015cf34c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.483842] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 604.488212] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03d6ff90-3512-43cf-a8d2-88975a065d86 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.504271] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 604.514187] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 604.514187] env[65107]: value = "task-5102162" [ 604.514187] env[65107]: _type = "Task" [ 604.514187] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.523616] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102162, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.541352] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 604.692360] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.693157] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 604.705881] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.750156] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102160, 'name': CreateVM_Task, 'duration_secs': 1.409635} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.750359] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 604.751012] env[65107]: WARNING neutronclient.v2_0.client [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.751491] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.751648] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.752021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 604.752639] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba7746df-93e2-4596-a48d-98bd8fef25be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.758497] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 604.758497] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523a193f-c07c-e760-5dfb-b60fee1c1a0f" [ 604.758497] env[65107]: _type = "Task" [ 604.758497] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.767377] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523a193f-c07c-e760-5dfb-b60fee1c1a0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.862179] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 604.875207] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance eac4b8f9-9cd5-44a3-a12e-9ec22767b907 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 604.875448] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 604.875575] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 4e70aaf3-5f1c-4a61-a790-a0a5e409170d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 604.892239] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 604.892499] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 604.892650] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 604.892896] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 604.893090] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 604.893278] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 604.893527] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.893698] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 604.893911] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 604.894212] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 604.894418] env[65107]: DEBUG nova.virt.hardware [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 604.895529] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ba16c2-c39e-4818-baa1-30e7e73aa506 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.905605] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16d0b21-ce3a-4946-b5e9-300dcc5ec889 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.026421] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102162, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.030060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.066676] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 605.134436] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 605.134672] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 605.190951] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Downloaded image file data e439aeac-2298-462c-be63-8218195135cf to vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk on the data store datastore2 {{(pid=65107) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 605.193117] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 605.193717] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Copying Virtual Disk [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk to [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 605.194815] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-125a8b65-787a-432e-b955-f7f2ddd1f405 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.204176] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 605.204176] env[65107]: value = "task-5102163" [ 605.204176] env[65107]: _type = "Task" [ 605.204176] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.208289] env[65107]: DEBUG nova.network.neutron [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 605.219192] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.272481] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 605.272748] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.272992] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.327777] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.327777] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.379201] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ccd68268-d0fc-406e-896f-c61b114dd75d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 605.379378] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b779b5ad-7d93-47c2-b824-6d76246c00f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 605.379507] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e3510c31-d6be-4e9f-a0a2-a662123861e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 605.379741] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 31195587-86ce-4379-946d-1abb51c35f08 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 605.525589] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102162, 'name': ReconfigVM_Task, 'duration_secs': 0.54553} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.526222] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Reconfigured VM instance instance-00000002 to attach disk [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 605.526984] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8035f44b-96a2-4b81-b54c-038625e060ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.534950] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 605.534950] env[65107]: value = "task-5102164" [ 605.534950] env[65107]: _type = "Task" [ 605.534950] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.545962] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102164, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.612650] env[65107]: DEBUG nova.network.neutron [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Successfully updated port: 74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 605.667672] env[65107]: WARNING neutronclient.v2_0.client [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 605.668087] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.668177] env[65107]: WARNING openstack [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.716562] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.886197] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 8306ff90-b8bd-4270-8133-96abe483156b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 605.886677] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 605.886677] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=100GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '7', 'num_vm_active': '1', 'num_task_None': '2', 'num_os_type_None': '7', 'num_proj_1134e7eca4244c8b8b038a72efa4210f': '1', 'io_workload': '6', 'num_vm_building': '6', 'num_task_spawning': '5', 'num_proj_a2d4844feda0432ebe1ef160b0070b83': '2', 'num_proj_aaf15dff8ae0472d94c9fecd0ba87a20': '1', 'num_proj_4df06b84d1cd45b28b8ea33565ac45f1': '1', 'num_proj_2882c2c7949045519a146655bb694a3f': '1', 'num_proj_1ae57a7184354c6b97c25d47135bcd40': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 605.990740] env[65107]: DEBUG nova.network.neutron [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance_info_cache with network_info: [{"id": "086361c0-3cd7-4389-b5f0-c550a2adc834", "address": "fa:16:3e:bf:a0:a1", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086361c0-3c", "ovs_interfaceid": "086361c0-3cd7-4389-b5f0-c550a2adc834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 606.051601] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102164, 'name': Rename_Task, 'duration_secs': 0.261736} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.051862] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 606.052137] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02d55e4e-ec70-4552-861c-834493de439b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.060662] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 606.060662] env[65107]: value = "task-5102165" [ 606.060662] env[65107]: _type = "Task" [ 606.060662] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.072951] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102165, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.110967] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a6b5f4-5f9b-42cc-b954-63243c4f2e6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.114366] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "refresh_cache-0ef89aea-5373-4d07-9587-37890c08d35b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.114563] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquired lock "refresh_cache-0ef89aea-5373-4d07-9587-37890c08d35b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.114743] env[65107]: DEBUG nova.network.neutron [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 606.123607] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17419da-d9d8-4389-ac7f-9ac34c5e614e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.162394] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5a76d9-038e-4f89-8e38-517ddb78f733 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.172279] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc853bd-2268-48af-bb6b-e7d76c1d59af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.188856] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.216595] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102163, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.307441] env[65107]: INFO nova.compute.manager [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Rebuilding instance [ 606.368449] env[65107]: DEBUG nova.compute.manager [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 606.369850] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908c91cf-c2af-49bc-929b-ba126cd0c3e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.494701] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 606.495126] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Instance network_info: |[{"id": "086361c0-3cd7-4389-b5f0-c550a2adc834", "address": "fa:16:3e:bf:a0:a1", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086361c0-3c", "ovs_interfaceid": "086361c0-3cd7-4389-b5f0-c550a2adc834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 606.495617] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:a0:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '086361c0-3cd7-4389-b5f0-c550a2adc834', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 606.504354] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Creating folder: Project (2882c2c7949045519a146655bb694a3f). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 606.504597] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68970365-b839-49d8-ac73-097fc590b7d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.517909] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Created folder: Project (2882c2c7949045519a146655bb694a3f) in parent group-v992574. [ 606.518139] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Creating folder: Instances. Parent ref: group-v992588. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 606.518337] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1489627-30b8-4e85-8bbd-36663c991b3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.530833] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Created folder: Instances in parent group-v992588. [ 606.530833] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 606.530833] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 606.530833] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ad2386a-6181-47b5-80f0-f2ed6cd5dc03 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.552921] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 606.552921] env[65107]: value = "task-5102168" [ 606.552921] env[65107]: _type = "Task" [ 606.552921] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.561473] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102168, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.573320] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102165, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.624487] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.625073] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.692288] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 606.726662] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102163, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.167198} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.727753] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Copied Virtual Disk [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk to [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 606.727947] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleting the datastore file [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf/tmp-sparse.vmdk {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 606.728238] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1a44bf4-b568-461f-b727-573af0f005ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.736463] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 606.736463] env[65107]: value = "task-5102169" [ 606.736463] env[65107]: _type = "Task" [ 606.736463] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.746169] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.754938] env[65107]: DEBUG nova.network.neutron [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 606.783383] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Received event network-changed-2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 606.783605] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Refreshing instance network info cache due to event network-changed-2686e3e0-c528-42bb-9444-050209faa6b1. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 606.783794] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "refresh_cache-e3510c31-d6be-4e9f-a0a2-a662123861e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.783941] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquired lock "refresh_cache-e3510c31-d6be-4e9f-a0a2-a662123861e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.787088] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Refreshing network info cache for port 2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 606.909569] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.910054] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.064763] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102168, 'name': CreateVM_Task, 'duration_secs': 0.487755} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.069686] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 607.069810] env[65107]: WARNING neutronclient.v2_0.client [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.070406] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.070613] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 607.071267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 607.072187] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e116b43d-87de-4b9d-800f-a004ff5c4755 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.080200] env[65107]: DEBUG oslo_vmware.api [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102165, 'name': PowerOnVM_Task, 'duration_secs': 0.924846} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.081198] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 607.081713] env[65107]: INFO nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Took 12.95 seconds to spawn the instance on the hypervisor. [ 607.082024] env[65107]: DEBUG nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 607.083459] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e046d846-4585-463b-8add-deb0496c34ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.090034] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 607.090034] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e7503b-e5bc-bf04-b978-5e9ff0a21f01" [ 607.090034] env[65107]: _type = "Task" [ 607.090034] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.107506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 607.107687] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 607.108113] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.188125] env[65107]: WARNING neutronclient.v2_0.client [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.188775] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.189126] env[65107]: WARNING openstack [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.202041] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 607.202432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.388s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 607.202545] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.659s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 607.204406] env[65107]: INFO nova.compute.claims [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.248942] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.047328} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.249331] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 607.249561] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Moving file from [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24/e439aeac-2298-462c-be63-8218195135cf to [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf. {{(pid=65107) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 607.250261] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-2e180b3b-83e1-4d14-9e61-7a22c932f006 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.257914] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 607.257914] env[65107]: value = "task-5102170" [ 607.257914] env[65107]: _type = "Task" [ 607.257914] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.267498] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102170, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.294723] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.295394] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.295562] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.390823] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 607.391167] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-190f2683-e0f2-4997-ab65-03687f104aef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.400816] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 607.400816] env[65107]: value = "task-5102171" [ 607.400816] env[65107]: _type = "Task" [ 607.400816] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.412273] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.610471] env[65107]: INFO nova.compute.manager [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Took 19.84 seconds to build instance. [ 607.768419] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102170, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.067867} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.768717] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] File moved {{(pid=65107) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 607.768915] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Cleaning up location [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24 {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 607.769088] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleting the datastore file [datastore2] vmware_temp/b7ac5f01-8d92-48a9-a65b-d91c32604e24 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 607.769340] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1f5e1e2-e3e9-48c4-8677-e4eff30e6c83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.776406] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 607.776406] env[65107]: value = "task-5102172" [ 607.776406] env[65107]: _type = "Task" [ 607.776406] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.788863] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.915015] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102171, 'name': PowerOffVM_Task, 'duration_secs': 0.270261} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.915503] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 607.916297] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.917464] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9f0d60-3b6a-47de-a185-481593b373ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.927332] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 607.927614] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80af794a-9abd-49ed-8ddc-76c60cab0775 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.954979] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 607.955240] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 607.955476] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleting the datastore file [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 607.955686] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d74d998-9deb-4bf1-845e-f83b86dc4af2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.967416] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 607.967416] env[65107]: value = "task-5102174" [ 607.967416] env[65107]: _type = "Task" [ 607.967416] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.979971] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.008077] env[65107]: DEBUG nova.network.neutron [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Updating instance_info_cache with network_info: [{"id": "74eb113e-24ab-43aa-bb1c-e5427ac59a2a", "address": "fa:16:3e:14:de:c4", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74eb113e-24", "ovs_interfaceid": "74eb113e-24ab-43aa-bb1c-e5427ac59a2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 608.090901] env[65107]: DEBUG nova.compute.manager [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Received event network-changed-d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 608.090901] env[65107]: DEBUG nova.compute.manager [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Refreshing instance network info cache due to event network-changed-d2435e56-5f0a-44e9-8154-2f2dd4103c55. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 608.090901] env[65107]: DEBUG oslo_concurrency.lockutils [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Acquiring lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.091151] env[65107]: DEBUG oslo_concurrency.lockutils [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Acquired lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 608.091696] env[65107]: DEBUG nova.network.neutron [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Refreshing network info cache for port d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 608.115680] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1098d80d-3092-4763-875f-f60296eaeeb9 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.354s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 608.249737] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.250142] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.294802] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031155} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.295172] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 608.296046] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb38c014-9964-4ea8-9d65-47c6893feb7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.305960] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 608.305960] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529957d5-2b6d-cd76-b79a-e4657eaab0ba" [ 608.305960] env[65107]: _type = "Task" [ 608.305960] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.319036] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529957d5-2b6d-cd76-b79a-e4657eaab0ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.437667] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81d0a96-5582-4e2e-9470-bbf930a38563 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.448451] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4859ea-7828-4f90-bad7-c8f82b95453e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.485633] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4c9dd1-2b39-4afa-8f7a-cf584e7e9abe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.495616] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296697} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.498594] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 608.498594] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 608.498594] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.502061] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d7643c-e26d-4f8b-a173-b8e3f1721f2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.523756] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Releasing lock "refresh_cache-0ef89aea-5373-4d07-9587-37890c08d35b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 608.523756] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Instance network_info: |[{"id": "74eb113e-24ab-43aa-bb1c-e5427ac59a2a", "address": "fa:16:3e:14:de:c4", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74eb113e-24", "ovs_interfaceid": "74eb113e-24ab-43aa-bb1c-e5427ac59a2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 608.523988] env[65107]: DEBUG nova.compute.provider_tree [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 608.523988] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:de:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74eb113e-24ab-43aa-bb1c-e5427ac59a2a', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 608.530825] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Creating folder: Project (1ae57a7184354c6b97c25d47135bcd40). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 608.531222] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5f722a7-cde9-4b28-9548-a396e0d3d626 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.544882] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Created folder: Project (1ae57a7184354c6b97c25d47135bcd40) in parent group-v992574. [ 608.545136] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Creating folder: Instances. Parent ref: group-v992591. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 608.545394] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a56b5134-2f79-4e8d-90f5-cf3cedb2f386 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.557511] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Created folder: Instances in parent group-v992591. [ 608.557702] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 608.558024] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 608.558733] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-680ed010-034f-48cf-9aeb-f30bb800beb9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.584638] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 608.584638] env[65107]: value = "task-5102177" [ 608.584638] env[65107]: _type = "Task" [ 608.584638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.593996] env[65107]: WARNING neutronclient.v2_0.client [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 608.594570] env[65107]: WARNING openstack [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.594896] env[65107]: WARNING openstack [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.607629] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102177, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.818257] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529957d5-2b6d-cd76-b79a-e4657eaab0ba, 'name': SearchDatastore_Task, 'duration_secs': 0.025464} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.818583] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 608.818977] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 4e70aaf3-5f1c-4a61-a790-a0a5e409170d/4e70aaf3-5f1c-4a61-a790-a0a5e409170d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 608.819419] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 608.819513] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 608.819761] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99d7781c-a42a-4680-b420-23bc5d0b2afb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.822047] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9c9375b-37c2-4a7d-a052-c2feb94663f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.830444] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 608.830444] env[65107]: value = "task-5102178" [ 608.830444] env[65107]: _type = "Task" [ 608.830444] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.835995] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 608.836299] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 608.838377] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d4220ab-f0d9-49d6-8a5d-a18b605725d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.844936] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102178, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.850679] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 608.850679] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522efd6b-a51a-fa78-2d7f-fe7438307fbc" [ 608.850679] env[65107]: _type = "Task" [ 608.850679] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.863141] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522efd6b-a51a-fa78-2d7f-fe7438307fbc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.934387] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 608.934387] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.934387] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.060080] env[65107]: ERROR nova.scheduler.client.report [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [req-e886609d-d3ee-42d5-b395-7bacafce411c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e886609d-d3ee-42d5-b395-7bacafce411c"}]} [ 609.083245] env[65107]: DEBUG nova.scheduler.client.report [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 609.103323] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102177, 'name': CreateVM_Task, 'duration_secs': 0.413532} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.103323] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 609.103323] env[65107]: WARNING neutronclient.v2_0.client [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.103849] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.104054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.104972] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 609.104972] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6733196b-4965-46ab-9dfd-25d35f13550a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.109224] env[65107]: DEBUG nova.scheduler.client.report [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 609.109379] env[65107]: DEBUG nova.compute.provider_tree [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 609.117261] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 609.117261] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528244c7-e7ae-4ff5-639f-b4ac6b0fb5b7" [ 609.117261] env[65107]: _type = "Task" [ 609.117261] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.131373] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528244c7-e7ae-4ff5-639f-b4ac6b0fb5b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.132953] env[65107]: DEBUG nova.scheduler.client.report [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 609.168361] env[65107]: DEBUG nova.scheduler.client.report [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 609.230478] env[65107]: WARNING openstack [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.230881] env[65107]: WARNING openstack [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.346757] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102178, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.367240] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522efd6b-a51a-fa78-2d7f-fe7438307fbc, 'name': SearchDatastore_Task, 'duration_secs': 0.021474} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.368457] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e851553-035d-4a26-99ea-55a2d104aa58 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.380314] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 609.380314] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ad371f-2868-bb8f-5161-1d3b022549a8" [ 609.380314] env[65107]: _type = "Task" [ 609.380314] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.386904] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ac2ffd-621a-4ce5-b016-65de1b459a09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.392851] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ad371f-2868-bb8f-5161-1d3b022549a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.399082] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47a03bb-58a0-4416-a2ba-b204829134ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.433118] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa853dfc-dc6d-4450-9753-c2fbfeada9c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.442820] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c3897c-920b-4293-b1c9-b96dfc91b988 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.458703] env[65107]: DEBUG nova.compute.provider_tree [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 609.548146] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Updated VIF entry in instance network info cache for port 2686e3e0-c528-42bb-9444-050209faa6b1. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 609.548508] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Updating instance_info_cache with network_info: [{"id": "2686e3e0-c528-42bb-9444-050209faa6b1", "address": "fa:16:3e:89:7b:5d", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2686e3e0-c5", "ovs_interfaceid": "2686e3e0-c528-42bb-9444-050209faa6b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 609.553230] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 609.554098] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 609.554098] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 609.554098] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 609.554098] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 609.554098] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 609.554322] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.554423] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 609.554584] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 609.554742] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 609.554910] env[65107]: DEBUG nova.virt.hardware [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 609.555804] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5d327d-6bf4-4d4f-b840-abd1e2c0bc00 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.565082] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da54312-fd24-498f-96af-21c94a4017ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.581278] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 609.587632] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 609.587877] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 609.588113] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5aea410-4e08-4ea4-a411-b415ec2b9d0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.608029] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 609.608029] env[65107]: value = "task-5102179" [ 609.608029] env[65107]: _type = "Task" [ 609.608029] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.617872] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102179, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.627165] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528244c7-e7ae-4ff5-639f-b4ac6b0fb5b7, 'name': SearchDatastore_Task, 'duration_secs': 0.066949} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.627481] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.627707] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 609.627911] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.842362] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102178, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625826} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.842803] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 4e70aaf3-5f1c-4a61-a790-a0a5e409170d/4e70aaf3-5f1c-4a61-a790-a0a5e409170d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 609.843073] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 609.843821] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdd6db66-d62d-47f0-bcac-cc43f11b87e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.852565] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 609.852565] env[65107]: value = "task-5102180" [ 609.852565] env[65107]: _type = "Task" [ 609.852565] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.868680] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102180, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.890919] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ad371f-2868-bb8f-5161-1d3b022549a8, 'name': SearchDatastore_Task, 'duration_secs': 0.053627} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.890919] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.891211] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 31195587-86ce-4379-946d-1abb51c35f08/31195587-86ce-4379-946d-1abb51c35f08.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 609.891494] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.891893] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 609.893017] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f79bdb00-d8ee-477d-8e67-c14ddb4505fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.894815] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c23f7c51-beca-46e6-a854-b0b01da42665 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.902322] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 609.902322] env[65107]: value = "task-5102181" [ 609.902322] env[65107]: _type = "Task" [ 609.902322] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.907634] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 609.907950] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 609.909174] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02782654-a1fc-44a3-a637-287b947c72eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.915070] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.919594] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 609.919594] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5235a9fe-771f-2c47-5c7d-a2f50b35a0cf" [ 609.919594] env[65107]: _type = "Task" [ 609.919594] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.929736] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5235a9fe-771f-2c47-5c7d-a2f50b35a0cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.002057] env[65107]: DEBUG nova.scheduler.client.report [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 15 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 610.002391] env[65107]: DEBUG nova.compute.provider_tree [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 15 to 16 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 610.002572] env[65107]: DEBUG nova.compute.provider_tree [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 610.050998] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Releasing lock "refresh_cache-e3510c31-d6be-4e9f-a0a2-a662123861e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 610.051654] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Received event network-vif-plugged-7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 610.051869] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "31195587-86ce-4379-946d-1abb51c35f08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.052095] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Lock "31195587-86ce-4379-946d-1abb51c35f08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.052267] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Lock "31195587-86ce-4379-946d-1abb51c35f08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 610.052434] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] No waiting events found dispatching network-vif-plugged-7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 610.052599] env[65107]: WARNING nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Received unexpected event network-vif-plugged-7518822c-68c7-4bd3-8296-e61bd76671e2 for instance with vm_state building and task_state spawning. [ 610.052764] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Received event network-changed-7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 610.052997] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Refreshing instance network info cache due to event network-changed-7518822c-68c7-4bd3-8296-e61bd76671e2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 610.053204] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "refresh_cache-31195587-86ce-4379-946d-1abb51c35f08" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.053340] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquired lock "refresh_cache-31195587-86ce-4379-946d-1abb51c35f08" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.053656] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Refreshing network info cache for port 7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 610.118238] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102179, 'name': CreateVM_Task, 'duration_secs': 0.330822} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.118423] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 610.118812] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.118984] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.119336] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 610.119596] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fabeead1-f7a6-4721-9279-5c5a53f0a97d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.125237] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 610.125237] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525ffd25-9d83-3268-c07d-aaec3de1c9cc" [ 610.125237] env[65107]: _type = "Task" [ 610.125237] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.135058] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525ffd25-9d83-3268-c07d-aaec3de1c9cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.317935] env[65107]: WARNING neutronclient.v2_0.client [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 610.318676] env[65107]: WARNING openstack [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.319022] env[65107]: WARNING openstack [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.363424] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102180, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073945} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.363713] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 610.364617] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf2f306-0170-4804-b579-ef8aadfff4c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.395331] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] 4e70aaf3-5f1c-4a61-a790-a0a5e409170d/4e70aaf3-5f1c-4a61-a790-a0a5e409170d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 610.396657] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be32d32b-7e25-47f2-9e4a-092b3c1e8a59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.421452] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.426595] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 610.426595] env[65107]: value = "task-5102182" [ 610.426595] env[65107]: _type = "Task" [ 610.426595] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.433064] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5235a9fe-771f-2c47-5c7d-a2f50b35a0cf, 'name': SearchDatastore_Task, 'duration_secs': 0.021378} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.435169] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-325c7bed-0454-4108-a79c-773e0f75b6f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.441704] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102182, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.445812] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 610.445812] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523f0314-4d2d-2c2c-361b-024bf1ff987c" [ 610.445812] env[65107]: _type = "Task" [ 610.445812] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.455575] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523f0314-4d2d-2c2c-361b-024bf1ff987c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.507710] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.305s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 610.508433] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 610.511472] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.806s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.513183] env[65107]: INFO nova.compute.claims [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.557355] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 610.558947] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.559108] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.638620] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525ffd25-9d83-3268-c07d-aaec3de1c9cc, 'name': SearchDatastore_Task, 'duration_secs': 0.03728} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.639054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 610.639322] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 610.639584] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.639754] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.639944] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 610.640273] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc1ec02f-4d2d-4ad8-abd9-18444bcdbf27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.653427] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 610.653753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 610.654926] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-943a93ad-2693-4c1f-83db-08c6056be706 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.663464] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 610.663464] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524fe654-6109-9371-3468-1bb041106d03" [ 610.663464] env[65107]: _type = "Task" [ 610.663464] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.675354] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524fe654-6109-9371-3468-1bb041106d03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.926475] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102181, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.938569] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102182, 'name': ReconfigVM_Task, 'duration_secs': 0.309951} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.939706] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Reconfigured VM instance instance-00000004 to attach disk [datastore2] 4e70aaf3-5f1c-4a61-a790-a0a5e409170d/4e70aaf3-5f1c-4a61-a790-a0a5e409170d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 610.939868] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-632441e0-e56c-4346-a6c4-673beee9663e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.951758] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 610.951758] env[65107]: value = "task-5102183" [ 610.951758] env[65107]: _type = "Task" [ 610.951758] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.960023] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523f0314-4d2d-2c2c-361b-024bf1ff987c, 'name': SearchDatastore_Task, 'duration_secs': 0.019235} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.960726] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 610.961051] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b779b5ad-7d93-47c2-b824-6d76246c00f5/b779b5ad-7d93-47c2-b824-6d76246c00f5.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 610.961354] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.961545] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 610.961768] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-545adfa1-f57c-4c52-8373-0b80f005d9ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.966699] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3115af6d-4f92-4f42-9821-458f033b0a89 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.971914] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102183, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.982035] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 610.982035] env[65107]: value = "task-5102184" [ 610.982035] env[65107]: _type = "Task" [ 610.982035] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.992720] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.994258] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 610.994449] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 610.995305] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a5526d5-f9b3-4030-884b-99993b61b40d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.004764] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 611.004764] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529d4472-6835-9b13-85f2-a50477c705d1" [ 611.004764] env[65107]: _type = "Task" [ 611.004764] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.012948] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529d4472-6835-9b13-85f2-a50477c705d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.017813] env[65107]: DEBUG nova.compute.utils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 611.022017] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 611.022235] env[65107]: DEBUG nova.network.neutron [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 611.022535] env[65107]: WARNING neutronclient.v2_0.client [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.022960] env[65107]: WARNING neutronclient.v2_0.client [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.024061] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.026408] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.177278] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524fe654-6109-9371-3468-1bb041106d03, 'name': SearchDatastore_Task, 'duration_secs': 0.026501} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.178265] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bae68a2-d32c-4850-9997-db0d8981ceda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.184124] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 611.184124] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f1cb16-de5a-2178-85f1-72c612095a88" [ 611.184124] env[65107]: _type = "Task" [ 611.184124] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.192874] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f1cb16-de5a-2178-85f1-72c612095a88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.425049] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102181, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.12349} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.425049] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 31195587-86ce-4379-946d-1abb51c35f08/31195587-86ce-4379-946d-1abb51c35f08.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 611.425049] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 611.425372] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c23e030-2e70-4e60-9c73-ed38c8af9c58 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.433139] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 611.433139] env[65107]: value = "task-5102185" [ 611.433139] env[65107]: _type = "Task" [ 611.433139] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.443893] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102185, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.465582] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102183, 'name': Rename_Task, 'duration_secs': 0.165236} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.465992] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 611.466358] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df1744c4-5059-45e8-bf09-e4dacd832caa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.476479] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 611.476479] env[65107]: value = "task-5102186" [ 611.476479] env[65107]: _type = "Task" [ 611.476479] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.489864] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.498059] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102184, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.516994] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529d4472-6835-9b13-85f2-a50477c705d1, 'name': SearchDatastore_Task, 'duration_secs': 0.013236} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.518031] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2d483f1-2e60-4936-bb80-4dac45f17019 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.522030] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 611.529510] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 611.529510] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52500da0-a9d5-14af-9c81-31803ab8ed84" [ 611.529510] env[65107]: _type = "Task" [ 611.529510] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.545170] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52500da0-a9d5-14af-9c81-31803ab8ed84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.648341] env[65107]: DEBUG nova.policy [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b7aff72042e4b218d05fe6251f87adb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd98fae2faa0a40c798e251737ccb31c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 611.701273] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f1cb16-de5a-2178-85f1-72c612095a88, 'name': SearchDatastore_Task, 'duration_secs': 0.034635} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.701605] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 611.701869] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 611.702170] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed36011a-85f9-45cd-b22f-acc8cdc4246f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.715799] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 611.715799] env[65107]: value = "task-5102187" [ 611.715799] env[65107]: _type = "Task" [ 611.715799] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.728747] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.790362] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357cc977-6b95-401e-9ddc-3820f392ec4d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.800177] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575a8219-50c1-4e6c-92ec-0c4c7e2e6fa9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.840811] env[65107]: DEBUG nova.network.neutron [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Updated VIF entry in instance network info cache for port d2435e56-5f0a-44e9-8154-2f2dd4103c55. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 611.841399] env[65107]: DEBUG nova.network.neutron [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Updating instance_info_cache with network_info: [{"id": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "address": "fa:16:3e:ae:a6:0b", "network": {"id": "4f3c2189-942b-47c1-95bd-5064d80e3a76", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-78304541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4df06b84d1cd45b28b8ea33565ac45f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0d2101e-2d93-4310-a242-af2d9ecdaf9b", "external-id": "nsx-vlan-transportzone-121", "segmentation_id": 121, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2435e56-5f", "ovs_interfaceid": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 611.846756] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2905d4a6-5542-4545-9ffc-1839e10ac07d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.853268] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f39e15-af27-4d79-a377-3030e4c66cb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.871484] env[65107]: DEBUG nova.compute.provider_tree [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.943609] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102185, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091952} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.943860] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 611.944684] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477f48b8-da6d-40d1-b78e-ebcb1207f69a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.968579] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 31195587-86ce-4379-946d-1abb51c35f08/31195587-86ce-4379-946d-1abb51c35f08.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 611.969095] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f791b5da-62c2-4222-bcb4-8c07c50b397d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.996542] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.833551} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.001102] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b779b5ad-7d93-47c2-b824-6d76246c00f5/b779b5ad-7d93-47c2-b824-6d76246c00f5.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 612.001378] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 612.001701] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 612.001701] env[65107]: value = "task-5102188" [ 612.001701] env[65107]: _type = "Task" [ 612.001701] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.001920] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102186, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.002120] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5501f5c6-34f5-4c41-8e99-a1a1679505fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.017096] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.018767] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 612.018767] env[65107]: value = "task-5102189" [ 612.018767] env[65107]: _type = "Task" [ 612.018767] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.035694] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102189, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.046335] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52500da0-a9d5-14af-9c81-31803ab8ed84, 'name': SearchDatastore_Task, 'duration_secs': 0.063728} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.046740] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.047077] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311/fdb5603d-24b0-4fdd-b8cc-25c0ebee5311.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 612.047322] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.047500] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 612.047721] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fa75a48-656e-4c94-b89a-813a59d8ceb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.050285] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b1624e8-ed0e-47a0-a9d6-c23931149b27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.062283] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 612.062283] env[65107]: value = "task-5102190" [ 612.062283] env[65107]: _type = "Task" [ 612.062283] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.067805] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.068055] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 612.069436] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea18c14b-1dce-49bd-b112-d2cdf671e22f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.076613] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.080615] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 612.080615] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ef42a1-f4fe-77a3-f348-ed5b85fbd9bd" [ 612.080615] env[65107]: _type = "Task" [ 612.080615] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.091802] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ef42a1-f4fe-77a3-f348-ed5b85fbd9bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.182901] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.183198] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.228232] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102187, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.348379] env[65107]: DEBUG oslo_concurrency.lockutils [req-4578a0bd-24d1-4867-a972-d18780d11ec8 req-52624888-066a-4eef-9d36-f336e3537e5c service nova] Releasing lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.378790] env[65107]: DEBUG nova.scheduler.client.report [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 612.498026] env[65107]: DEBUG oslo_vmware.api [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102186, 'name': PowerOnVM_Task, 'duration_secs': 0.592744} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.498026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 612.498026] env[65107]: INFO nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Took 13.95 seconds to spawn the instance on the hypervisor. [ 612.498026] env[65107]: DEBUG nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 612.498026] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1adfecf-0830-4211-a8a3-f6320357b730 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.520924] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.531898] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102189, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086742} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.532504] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 612.533487] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d838936e-3411-4313-86ee-e11247beaa3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.537517] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 612.571148] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] b779b5ad-7d93-47c2-b824-6d76246c00f5/b779b5ad-7d93-47c2-b824-6d76246c00f5.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 612.573529] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20f58618-97e2-4b17-a127-bf8466d3818c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.601940] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 612.602431] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 612.603192] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 612.603192] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 612.603192] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 612.603335] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 612.603556] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.606057] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 612.606057] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 612.606057] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 612.606057] env[65107]: DEBUG nova.virt.hardware [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 612.606057] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c74601-2a6a-46d6-85c1-b4156f6da07f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.617720] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 612.617720] env[65107]: value = "task-5102191" [ 612.617720] env[65107]: _type = "Task" [ 612.617720] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.618355] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ef42a1-f4fe-77a3-f348-ed5b85fbd9bd, 'name': SearchDatastore_Task, 'duration_secs': 0.019387} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.618594] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.623736] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-048d2042-b616-4fe5-b46b-ab93c2b801ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.634367] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.639107] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 612.639107] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52305282-fdff-79df-a6da-2ba28c4af79f" [ 612.639107] env[65107]: _type = "Task" [ 612.639107] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.640030] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ece1c2-08d2-4c0a-b460-a3c32060e0a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.656735] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52305282-fdff-79df-a6da-2ba28c4af79f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.728510] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102187, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.883461] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.883995] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 612.889146] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.859s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.892345] env[65107]: INFO nova.compute.claims [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.920878] env[65107]: DEBUG nova.network.neutron [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Successfully created port: e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 613.026650] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.029275] env[65107]: INFO nova.compute.manager [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Took 23.77 seconds to build instance. [ 613.089213] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102190, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.133789] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102191, 'name': ReconfigVM_Task, 'duration_secs': 0.476024} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.134106] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Reconfigured VM instance instance-00000005 to attach disk [datastore2] b779b5ad-7d93-47c2-b824-6d76246c00f5/b779b5ad-7d93-47c2-b824-6d76246c00f5.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 613.135308] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19e79aac-acec-41cc-9402-2a87bfb08a84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.140855] env[65107]: DEBUG nova.compute.manager [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Received event network-vif-plugged-74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 613.141083] env[65107]: DEBUG oslo_concurrency.lockutils [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Acquiring lock "0ef89aea-5373-4d07-9587-37890c08d35b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.141295] env[65107]: DEBUG oslo_concurrency.lockutils [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Lock "0ef89aea-5373-4d07-9587-37890c08d35b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.141480] env[65107]: DEBUG oslo_concurrency.lockutils [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Lock "0ef89aea-5373-4d07-9587-37890c08d35b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.142250] env[65107]: DEBUG nova.compute.manager [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] No waiting events found dispatching network-vif-plugged-74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 613.142250] env[65107]: WARNING nova.compute.manager [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Received unexpected event network-vif-plugged-74eb113e-24ab-43aa-bb1c-e5427ac59a2a for instance with vm_state building and task_state spawning. [ 613.142250] env[65107]: DEBUG nova.compute.manager [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Received event network-changed-74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 613.142250] env[65107]: DEBUG nova.compute.manager [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Refreshing instance network info cache due to event network-changed-74eb113e-24ab-43aa-bb1c-e5427ac59a2a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 613.142453] env[65107]: DEBUG oslo_concurrency.lockutils [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Acquiring lock "refresh_cache-0ef89aea-5373-4d07-9587-37890c08d35b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.143844] env[65107]: DEBUG oslo_concurrency.lockutils [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Acquired lock "refresh_cache-0ef89aea-5373-4d07-9587-37890c08d35b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.143844] env[65107]: DEBUG nova.network.neutron [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Refreshing network info cache for port 74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 613.151478] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 613.151478] env[65107]: value = "task-5102192" [ 613.151478] env[65107]: _type = "Task" [ 613.151478] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.166201] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52305282-fdff-79df-a6da-2ba28c4af79f, 'name': SearchDatastore_Task, 'duration_secs': 0.038596} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.166836] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 613.167348] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0ef89aea-5373-4d07-9587-37890c08d35b/0ef89aea-5373-4d07-9587-37890c08d35b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 613.167530] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c99a0fc2-02f1-4188-ad27-148f366b1450 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.174965] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102192, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.182205] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 613.182205] env[65107]: value = "task-5102193" [ 613.182205] env[65107]: _type = "Task" [ 613.182205] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.191719] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102193, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.228306] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102187, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.117416} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.229128] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 613.229128] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 613.229128] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9463e052-ccee-4bd3-90b8-1e2a2686c77a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.238404] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 613.238404] env[65107]: value = "task-5102194" [ 613.238404] env[65107]: _type = "Task" [ 613.238404] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.248894] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102194, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.401271] env[65107]: DEBUG nova.compute.utils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 613.406015] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 613.406015] env[65107]: DEBUG nova.network.neutron [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 613.406015] env[65107]: WARNING neutronclient.v2_0.client [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.406346] env[65107]: WARNING neutronclient.v2_0.client [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.407146] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.407509] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.522024] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102188, 'name': ReconfigVM_Task, 'duration_secs': 1.270832} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.522641] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 31195587-86ce-4379-946d-1abb51c35f08/31195587-86ce-4379-946d-1abb51c35f08.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 613.523353] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-787a42a4-cc04-4051-b93c-e15f124f4414 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.530951] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 613.530951] env[65107]: value = "task-5102195" [ 613.530951] env[65107]: _type = "Task" [ 613.530951] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.535541] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e93a6c6-a13f-447a-afcf-03641ea9a008 tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.321s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.541717] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102195, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.586062] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102190, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.136099} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.586514] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311/fdb5603d-24b0-4fdd-b8cc-25c0ebee5311.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 613.586780] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 613.587108] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-222e9bb8-0c8c-45a8-8666-9923092e7d5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.596871] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 613.596871] env[65107]: value = "task-5102196" [ 613.596871] env[65107]: _type = "Task" [ 613.596871] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.606295] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102196, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.652286] env[65107]: WARNING neutronclient.v2_0.client [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.653318] env[65107]: WARNING openstack [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.653832] env[65107]: WARNING openstack [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.672598] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102192, 'name': Rename_Task, 'duration_secs': 0.176193} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.672965] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 613.673245] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51a4ab1f-d34b-4efb-9746-6fc37b3c5eda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.684978] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 613.684978] env[65107]: value = "task-5102197" [ 613.684978] env[65107]: _type = "Task" [ 613.684978] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.710303] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102193, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.710574] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102197, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.752595] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102194, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086023} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.752595] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 613.752595] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046e4710-c369-42f8-aca3-ee10a4f3d610 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.777406] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 613.777528] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39f597cc-625e-42e1-9e91-306813e263a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.799863] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 613.799863] env[65107]: value = "task-5102198" [ 613.799863] env[65107]: _type = "Task" [ 613.799863] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.811971] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102198, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.883125] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.884080] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.884602] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.906571] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 614.042754] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102195, 'name': Rename_Task, 'duration_secs': 0.458701} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.043192] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 614.043496] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d83ff901-ac82-4d8f-af4d-82c0e650d773 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.053564] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 614.053564] env[65107]: value = "task-5102199" [ 614.053564] env[65107]: _type = "Task" [ 614.053564] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.066748] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.111993] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102196, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114511} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.111993] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.112616] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3470c511-89fe-4c16-88a7-8a84446a6fe7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.139670] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311/fdb5603d-24b0-4fdd-b8cc-25c0ebee5311.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.141286] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7d0f8dc-8e7a-4c24-a347-741885c7dca8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.157632] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1d3c90-9ec8-4011-912e-11a610e79cbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.167452] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2d0c40-c9bd-429c-8551-858362ed8a06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.172082] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 614.172082] env[65107]: value = "task-5102200" [ 614.172082] env[65107]: _type = "Task" [ 614.172082] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.211320] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c51eb00-4b54-4316-b624-1613332795ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.217889] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102200, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.231879] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102197, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.232220] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102193, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.768702} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.232534] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0ef89aea-5373-4d07-9587-37890c08d35b/0ef89aea-5373-4d07-9587-37890c08d35b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.233417] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.234161] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844eb5d6-5826-4765-a719-a05c51de9941 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.239405] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-689a1701-f32c-40ce-8467-2c5c939d0559 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.255545] env[65107]: DEBUG nova.compute.provider_tree [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.260159] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 614.260159] env[65107]: value = "task-5102201" [ 614.260159] env[65107]: _type = "Task" [ 614.260159] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.269084] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102201, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.313047] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.571263] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102199, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.643869] env[65107]: DEBUG nova.network.neutron [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Successfully updated port: e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 614.688032] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102200, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.719661] env[65107]: DEBUG oslo_vmware.api [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102197, 'name': PowerOnVM_Task, 'duration_secs': 0.642237} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.722646] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 614.723081] env[65107]: INFO nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Took 14.45 seconds to spawn the instance on the hypervisor. [ 614.723344] env[65107]: DEBUG nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 614.724385] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dcd0a7-b5c9-4c7f-ab04-fb13f7b6626f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.745432] env[65107]: DEBUG nova.policy [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7edab34f8fd8407dac88b42cbe4b2bb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ac04479a6164cf7b18b9ad304a0c9b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 614.765037] env[65107]: DEBUG nova.scheduler.client.report [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.776850] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102201, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088425} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.776850] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.777925] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205295d9-95b3-4131-8951-0c61b1802c91 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.808516] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 0ef89aea-5373-4d07-9587-37890c08d35b/0ef89aea-5373-4d07-9587-37890c08d35b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.810079] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3762003b-ffa3-4d2f-a746-fbbf2c56022f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.850712] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102198, 'name': ReconfigVM_Task, 'duration_secs': 0.995} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.850712] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Reconfigured VM instance instance-00000001 to attach disk [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 614.850712] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 614.850712] env[65107]: value = "task-5102202" [ 614.850712] env[65107]: _type = "Task" [ 614.850712] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.850712] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd8c105b-f7eb-49a4-a3c4-139721ad8412 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.865817] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102202, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.867737] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 614.867737] env[65107]: value = "task-5102203" [ 614.867737] env[65107]: _type = "Task" [ 614.867737] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.882752] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102203, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.917974] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 614.955324] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 614.955907] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 614.956265] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 614.957100] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 614.957358] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 614.957673] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 614.958109] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.958481] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 614.958748] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 614.959552] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 614.959552] env[65107]: DEBUG nova.virt.hardware [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 614.960485] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0192fcad-f301-4d68-b9da-8bce8836f05e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.974238] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2d074b-d9d7-49a5-93e5-364d1121bf56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.069949] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102199, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.147141] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.147414] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquired lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 615.147535] env[65107]: DEBUG nova.network.neutron [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 615.192385] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102200, 'name': ReconfigVM_Task, 'duration_secs': 0.847252} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.192508] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Reconfigured VM instance instance-00000006 to attach disk [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311/fdb5603d-24b0-4fdd-b8cc-25c0ebee5311.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 615.193277] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72077f48-bef5-4c18-838b-6a1be7af39f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.201622] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 615.201622] env[65107]: value = "task-5102204" [ 615.201622] env[65107]: _type = "Task" [ 615.201622] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.213633] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102204, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.245831] env[65107]: INFO nova.compute.manager [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Took 25.89 seconds to build instance. [ 615.271549] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.271549] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 615.275579] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.209s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.277278] env[65107]: INFO nova.compute.claims [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.364633] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.379164] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102203, 'name': Rename_Task, 'duration_secs': 0.269298} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.379446] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 615.379692] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57a6824c-5bce-4225-aa51-3dadd2a0cc86 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.389195] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 615.389195] env[65107]: value = "task-5102205" [ 615.389195] env[65107]: _type = "Task" [ 615.389195] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.404655] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.521070] env[65107]: DEBUG nova.network.neutron [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Successfully created port: 02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 615.565926] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Updated VIF entry in instance network info cache for port 7518822c-68c7-4bd3-8296-e61bd76671e2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 615.566055] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Updating instance_info_cache with network_info: [{"id": "7518822c-68c7-4bd3-8296-e61bd76671e2", "address": "fa:16:3e:60:c5:8a", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.217", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7518822c-68", "ovs_interfaceid": "7518822c-68c7-4bd3-8296-e61bd76671e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.571677] env[65107]: DEBUG oslo_vmware.api [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102199, 'name': PowerOnVM_Task, 'duration_secs': 1.0495} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.573917] env[65107]: WARNING openstack [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.573966] env[65107]: WARNING openstack [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.584078] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 615.584078] env[65107]: INFO nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Took 19.27 seconds to spawn the instance on the hypervisor. [ 615.584078] env[65107]: DEBUG nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 615.584078] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3208d8-fa8f-425f-ba28-4b1872d89c55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.658936] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.659694] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.721026] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102204, 'name': Rename_Task, 'duration_secs': 0.355958} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.721026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 615.721026] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5928f89-20c6-4b79-bb1e-2a5ae70dce64 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.729066] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 615.729066] env[65107]: value = "task-5102206" [ 615.729066] env[65107]: _type = "Task" [ 615.729066] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.744512] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.748452] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f800c330-731f-417a-92f9-c9f0303eb033 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.407s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.786426] env[65107]: DEBUG nova.compute.utils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 615.794032] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 615.794032] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 615.794032] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.794526] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.795866] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.795866] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.869035] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102202, 'name': ReconfigVM_Task, 'duration_secs': 0.68417} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.869035] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 0ef89aea-5373-4d07-9587-37890c08d35b/0ef89aea-5373-4d07-9587-37890c08d35b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 615.869035] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a84a718f-de64-4992-8875-c021f1f63d6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.876227] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 615.876227] env[65107]: value = "task-5102207" [ 615.876227] env[65107]: _type = "Task" [ 615.876227] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.890395] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102207, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.903790] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102205, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.006063] env[65107]: DEBUG nova.network.neutron [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 616.074100] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Releasing lock "refresh_cache-31195587-86ce-4379-946d-1abb51c35f08" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 616.074457] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Received event network-vif-plugged-3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 616.074662] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "b779b5ad-7d93-47c2-b824-6d76246c00f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.074908] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.075259] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.075340] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] No waiting events found dispatching network-vif-plugged-3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 616.075508] env[65107]: WARNING nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Received unexpected event network-vif-plugged-3ce4a5c1-5dae-42ba-919f-736e66893243 for instance with vm_state building and task_state spawning. [ 616.075681] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Received event network-changed-3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 616.075844] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Refreshing instance network info cache due to event network-changed-3ce4a5c1-5dae-42ba-919f-736e66893243. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 616.076503] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "refresh_cache-b779b5ad-7d93-47c2-b824-6d76246c00f5" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.076503] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquired lock "refresh_cache-b779b5ad-7d93-47c2-b824-6d76246c00f5" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 616.076503] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Refreshing network info cache for port 3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 616.108140] env[65107]: INFO nova.compute.manager [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Took 27.31 seconds to build instance. [ 616.182472] env[65107]: DEBUG nova.policy [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4b2c7e86b0747ff8832c61ee96e1f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24d8b15ee5034b33a4c741412bcc71e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 616.243157] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.300128] env[65107]: WARNING neutronclient.v2_0.client [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.302294] env[65107]: WARNING openstack [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.302966] env[65107]: WARNING openstack [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.317652] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 616.372157] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.373135] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.406165] env[65107]: DEBUG oslo_vmware.api [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102205, 'name': PowerOnVM_Task, 'duration_secs': 0.826928} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.409423] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.409673] env[65107]: DEBUG nova.compute.manager [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 616.409980] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102207, 'name': Rename_Task, 'duration_secs': 0.276232} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.414195] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d89859-a2f9-4964-b4f5-6d0f10b73673 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.417158] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.417740] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac120f2b-c95e-4380-aa00-6b199cfc3e98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.431403] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 616.431403] env[65107]: value = "task-5102208" [ 616.431403] env[65107]: _type = "Task" [ 616.431403] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.446027] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.540864] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41adc765-7304-4dc3-8193-3a8d6394d8cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.550603] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5aa4865-9cea-452d-ba4c-0ffa5b2a4fee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.586759] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.587477] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.587833] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.600449] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c4d53f-cb2b-4356-af48-ed72823f0d65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.610376] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962b7721-daeb-4575-bcf5-b0a052be8693 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.615765] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb3f7653-3042-447f-9dc6-fdcea94161b2 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "31195587-86ce-4379-946d-1abb51c35f08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.829s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.631308] env[65107]: DEBUG nova.compute.provider_tree [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.744198] env[65107]: DEBUG oslo_vmware.api [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102206, 'name': PowerOnVM_Task, 'duration_secs': 0.937569} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.744650] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.744784] env[65107]: INFO nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Took 14.21 seconds to spawn the instance on the hypervisor. [ 616.744862] env[65107]: DEBUG nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 616.745748] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ccd0f6-5e41-4991-b780-dd88f0b674d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.858624] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Successfully created port: d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 616.944929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.950486] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102208, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.134748] env[65107]: DEBUG nova.scheduler.client.report [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.267168] env[65107]: INFO nova.compute.manager [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Took 27.14 seconds to build instance. [ 617.283671] env[65107]: DEBUG nova.network.neutron [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Successfully updated port: 02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 617.329414] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 617.344743] env[65107]: DEBUG nova.network.neutron [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Updated VIF entry in instance network info cache for port 74eb113e-24ab-43aa-bb1c-e5427ac59a2a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 617.345138] env[65107]: DEBUG nova.network.neutron [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Updating instance_info_cache with network_info: [{"id": "74eb113e-24ab-43aa-bb1c-e5427ac59a2a", "address": "fa:16:3e:14:de:c4", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74eb113e-24", "ovs_interfaceid": "74eb113e-24ab-43aa-bb1c-e5427ac59a2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 617.368374] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 617.368622] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 617.368776] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 617.368979] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 617.370269] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 617.371584] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 617.371584] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.371584] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 617.372034] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 617.372034] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 617.372175] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 617.373463] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09691756-aa62-4ec3-9e1c-a3de7938ce9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.384186] env[65107]: WARNING neutronclient.v2_0.client [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.384850] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.385238] env[65107]: WARNING openstack [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.394199] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26a64f4-1ff7-405a-b60d-3abcd06f93e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.446184] env[65107]: DEBUG oslo_vmware.api [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102208, 'name': PowerOnVM_Task, 'duration_secs': 0.809954} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.446506] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.446646] env[65107]: INFO nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Took 12.58 seconds to spawn the instance on the hypervisor. [ 617.446823] env[65107]: DEBUG nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 617.447629] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669061b4-31d1-4d34-a846-2ca8bbba68b4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.644867] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.645496] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 617.648069] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.703s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.648561] env[65107]: DEBUG nova.objects.instance [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 617.768138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-859d736f-33a6-4382-862b-6f0e1180160a tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.655s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.786312] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.786312] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquired lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.786444] env[65107]: DEBUG nova.network.neutron [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 617.848254] env[65107]: DEBUG oslo_concurrency.lockutils [req-cb415256-aaed-4d0f-9487-f310a803496f req-5f41fd37-5000-45d6-bd70-0d9ae9ac2ee1 service nova] Releasing lock "refresh_cache-0ef89aea-5373-4d07-9587-37890c08d35b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.968522] env[65107]: INFO nova.compute.manager [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Took 27.01 seconds to build instance. [ 618.002312] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.002721] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.042593] env[65107]: DEBUG nova.network.neutron [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updating instance_info_cache with network_info: [{"id": "e41be772-a45d-410b-acaf-09d30f9afa9d", "address": "fa:16:3e:f7:84:8e", "network": {"id": "9a766a32-cd65-4a5e-b338-308b5e640847", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-2025951181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d98fae2faa0a40c798e251737ccb31c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41be772-a4", "ovs_interfaceid": "e41be772-a45d-410b-acaf-09d30f9afa9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 618.152292] env[65107]: DEBUG nova.compute.utils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 618.153851] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 618.154159] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 618.154551] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.154905] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.155518] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.155918] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.289447] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.289836] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.471463] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1419320e-3f79-4c8a-9e73-c94778eb2f56 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "0ef89aea-5373-4d07-9587-37890c08d35b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.526s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 618.489455] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "7a3009bf-54a2-4565-a1aa-1d19286a4810" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.489686] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.546688] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Releasing lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.547152] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Instance network_info: |[{"id": "e41be772-a45d-410b-acaf-09d30f9afa9d", "address": "fa:16:3e:f7:84:8e", "network": {"id": "9a766a32-cd65-4a5e-b338-308b5e640847", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-2025951181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d98fae2faa0a40c798e251737ccb31c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41be772-a4", "ovs_interfaceid": "e41be772-a45d-410b-acaf-09d30f9afa9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 618.547673] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:84:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31e77685-b4dd-4810-80ef-24115ea9ea62', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e41be772-a45d-410b-acaf-09d30f9afa9d', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 618.555727] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Creating folder: Project (d98fae2faa0a40c798e251737ccb31c6). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 618.556032] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2e545d0-2a1c-41d1-8507-0f8565010aa9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.569143] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Created folder: Project (d98fae2faa0a40c798e251737ccb31c6) in parent group-v992574. [ 618.569347] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Creating folder: Instances. Parent ref: group-v992595. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 618.569903] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d1e5e7d-4589-4c0b-af11-fd5a2e570216 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.579845] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Created folder: Instances in parent group-v992595. [ 618.580061] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 618.580250] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 618.580458] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-575925f6-5f49-4296-a832-fab97a0a3726 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.599711] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 618.599711] env[65107]: value = "task-5102211" [ 618.599711] env[65107]: _type = "Task" [ 618.599711] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.607549] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102211, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.625805] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Successfully updated port: d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 618.666888] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 618.670047] env[65107]: DEBUG oslo_concurrency.lockutils [None req-494e3c61-8a71-4143-a350-9d60bb3f5348 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 618.769994] env[65107]: DEBUG nova.policy [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4b2c7e86b0747ff8832c61ee96e1f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24d8b15ee5034b33a4c741412bcc71e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 618.779607] env[65107]: DEBUG nova.network.neutron [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 618.824451] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.825371] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.826019] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.994880] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 619.108065] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Successfully created port: b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 619.115029] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102211, 'name': CreateVM_Task, 'duration_secs': 0.423959} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.115140] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 619.115642] env[65107]: WARNING neutronclient.v2_0.client [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.116067] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.116265] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.116621] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 619.116936] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99e7d35a-44fb-49d8-93b3-fd5902f7c138 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.122498] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 619.122498] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529411a5-30a8-a39b-fcb7-13f6d4fdb2d0" [ 619.122498] env[65107]: _type = "Task" [ 619.122498] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.124944] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.125400] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.136029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "refresh_cache-abb9dc0e-ce18-49c8-a472-d39a3fcec887" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.136244] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "refresh_cache-abb9dc0e-ce18-49c8-a472-d39a3fcec887" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.136447] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 619.147368] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529411a5-30a8-a39b-fcb7-13f6d4fdb2d0, 'name': SearchDatastore_Task, 'duration_secs': 0.015923} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.148600] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.148998] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 619.149157] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.149391] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.149560] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 619.150138] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-970b8777-35f9-447a-8976-57fbda1edd5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.161151] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 619.161471] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 619.162163] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bdaa23d-1e3a-45ed-8c6f-a039e2b17e8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.168791] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 619.168791] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52946c5b-13ad-ee2b-bb8d-1ab314b844ce" [ 619.168791] env[65107]: _type = "Task" [ 619.168791] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.180646] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52946c5b-13ad-ee2b-bb8d-1ab314b844ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.515598] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.515598] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.516028] env[65107]: INFO nova.compute.claims [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.638924] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.639392] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.676868] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 619.685227] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52946c5b-13ad-ee2b-bb8d-1ab314b844ce, 'name': SearchDatastore_Task, 'duration_secs': 0.015348} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.686182] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ea3ca05-2fa8-42c7-8807-d21358a26181 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.692583] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 619.692583] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f28e6b-83a8-694d-b684-c7a6bc77127e" [ 619.692583] env[65107]: _type = "Task" [ 619.692583] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.701564] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f28e6b-83a8-694d-b684-c7a6bc77127e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.710703] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 619.711666] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 619.711666] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 619.711666] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 619.711666] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 619.711666] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 619.711945] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.711945] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 619.712175] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 619.712348] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 619.712514] env[65107]: DEBUG nova.virt.hardware [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 619.713398] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ca8f92-87db-415f-930a-94fe972c7bca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.721778] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2bbd33-0098-4a9a-ac94-c5ce6b0b47a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.745488] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Updated VIF entry in instance network info cache for port 3ce4a5c1-5dae-42ba-919f-736e66893243. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 619.745488] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Updating instance_info_cache with network_info: [{"id": "3ce4a5c1-5dae-42ba-919f-736e66893243", "address": "fa:16:3e:c0:0c:58", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ce4a5c1-5d", "ovs_interfaceid": "3ce4a5c1-5dae-42ba-919f-736e66893243", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 619.785322] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 619.815261] env[65107]: WARNING neutronclient.v2_0.client [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.815358] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.815855] env[65107]: WARNING openstack [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.888528] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.888957] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.959511] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "be518c1d-edd7-40ff-b7cc-3310885b07b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.959742] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.172146] env[65107]: DEBUG nova.network.neutron [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updating instance_info_cache with network_info: [{"id": "02108638-8e1e-447a-9b94-a45adfd32161", "address": "fa:16:3e:49:92:f3", "network": {"id": "f011d711-ac42-4e0b-bcd5-8f569242741c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-475912691-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac04479a6164cf7b18b9ad304a0c9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02108638-8e", "ovs_interfaceid": "02108638-8e1e-447a-9b94-a45adfd32161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 620.211012] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f28e6b-83a8-694d-b684-c7a6bc77127e, 'name': SearchDatastore_Task, 'duration_secs': 0.015628} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.211329] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.211583] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 8306ff90-b8bd-4270-8133-96abe483156b/8306ff90-b8bd-4270-8133-96abe483156b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 620.212294] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0f9bd22-9726-4cce-af30-b119b787424c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.221162] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 620.221162] env[65107]: value = "task-5102212" [ 620.221162] env[65107]: _type = "Task" [ 620.221162] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.231668] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.253626] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Releasing lock "refresh_cache-b779b5ad-7d93-47c2-b824-6d76246c00f5" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.255073] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Received event network-vif-plugged-086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 620.255073] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.255073] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.255073] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 620.255346] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] No waiting events found dispatching network-vif-plugged-086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 620.255384] env[65107]: WARNING nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Received unexpected event network-vif-plugged-086361c0-3cd7-4389-b5f0-c550a2adc834 for instance with vm_state building and task_state spawning. [ 620.255539] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Received event network-changed-086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 620.255702] env[65107]: DEBUG nova.compute.manager [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Refreshing instance network info cache due to event network-changed-086361c0-3cd7-4389-b5f0-c550a2adc834. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 620.255894] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquiring lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.256044] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Acquired lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.256836] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Refreshing network info cache for port 086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 620.334684] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.337064] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.337319] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.462912] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 620.675102] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Releasing lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.675454] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Instance network_info: |[{"id": "02108638-8e1e-447a-9b94-a45adfd32161", "address": "fa:16:3e:49:92:f3", "network": {"id": "f011d711-ac42-4e0b-bcd5-8f569242741c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-475912691-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac04479a6164cf7b18b9ad304a0c9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02108638-8e", "ovs_interfaceid": "02108638-8e1e-447a-9b94-a45adfd32161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 620.675916] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:92:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24727047-6358-4015-86c1-394ab07fb88f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02108638-8e1e-447a-9b94-a45adfd32161', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 620.683555] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Creating folder: Project (9ac04479a6164cf7b18b9ad304a0c9b8). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.686602] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dde3f5c2-0396-49ba-bb4f-7c3f3df3fdcd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.699801] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Created folder: Project (9ac04479a6164cf7b18b9ad304a0c9b8) in parent group-v992574. [ 620.700159] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Creating folder: Instances. Parent ref: group-v992598. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.703374] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bceea888-4c5f-4f40-9426-709fab88fece {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.715893] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Created folder: Instances in parent group-v992598. [ 620.716211] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 620.716427] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 620.716674] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccad1e12-d8a6-4b77-83f1-eba6dfdbed9c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.743963] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102212, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.745602] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 620.745602] env[65107]: value = "task-5102215" [ 620.745602] env[65107]: _type = "Task" [ 620.745602] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.759236] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.760107] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.760443] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.768125] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102215, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.770467] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec2037a-008b-4e37-ae76-fba78e34a483 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.781990] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a71ee44-0a8a-4165-b548-4c695a298bfb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.827274] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a31d8d7-68b3-493e-9788-df437dcd1491 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.842055] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7400973d-d157-49ce-82cd-613f7f228b2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.872382] env[65107]: DEBUG nova.compute.provider_tree [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.992565] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.024798] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.024798] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 621.030980] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Successfully updated port: b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 621.112963] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Updating instance_info_cache with network_info: [{"id": "d06bc826-4580-47a4-a3b4-841380ee155f", "address": "fa:16:3e:37:92:b6", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd06bc826-45", "ovs_interfaceid": "d06bc826-4580-47a4-a3b4-841380ee155f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 621.189406] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.189873] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.244096] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102212, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727088} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.244540] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 8306ff90-b8bd-4270-8133-96abe483156b/8306ff90-b8bd-4270-8133-96abe483156b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.244821] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 621.245195] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b11f0d3-a8cc-4064-8d83-5d05a4ef2d22 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.262777] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102215, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.265575] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 621.265575] env[65107]: value = "task-5102216" [ 621.265575] env[65107]: _type = "Task" [ 621.265575] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.272230] env[65107]: WARNING neutronclient.v2_0.client [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.272779] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.273098] env[65107]: WARNING openstack [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.296036] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.381043] env[65107]: DEBUG nova.scheduler.client.report [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 621.390511] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updated VIF entry in instance network info cache for port 086361c0-3cd7-4389-b5f0-c550a2adc834. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 621.391399] env[65107]: DEBUG nova.network.neutron [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance_info_cache with network_info: [{"id": "086361c0-3cd7-4389-b5f0-c550a2adc834", "address": "fa:16:3e:bf:a0:a1", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086361c0-3c", "ovs_interfaceid": "086361c0-3cd7-4389-b5f0-c550a2adc834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 621.516273] env[65107]: DEBUG nova.compute.manager [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Received event network-vif-plugged-e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 621.516509] env[65107]: DEBUG oslo_concurrency.lockutils [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.517864] env[65107]: DEBUG oslo_concurrency.lockutils [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Lock "8306ff90-b8bd-4270-8133-96abe483156b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 621.517864] env[65107]: DEBUG oslo_concurrency.lockutils [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Lock "8306ff90-b8bd-4270-8133-96abe483156b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 621.517864] env[65107]: DEBUG nova.compute.manager [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] No waiting events found dispatching network-vif-plugged-e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 621.518605] env[65107]: WARNING nova.compute.manager [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Received unexpected event network-vif-plugged-e41be772-a45d-410b-acaf-09d30f9afa9d for instance with vm_state building and task_state spawning. [ 621.518983] env[65107]: DEBUG nova.compute.manager [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Received event network-changed-e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 621.518983] env[65107]: DEBUG nova.compute.manager [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Refreshing instance network info cache due to event network-changed-e41be772-a45d-410b-acaf-09d30f9afa9d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 621.519334] env[65107]: DEBUG oslo_concurrency.lockutils [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Acquiring lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.519334] env[65107]: DEBUG oslo_concurrency.lockutils [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Acquired lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.519700] env[65107]: DEBUG nova.network.neutron [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Refreshing network info cache for port e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 621.527298] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 621.532871] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "refresh_cache-eac4b8f9-9cd5-44a3-a12e-9ec22767b907" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.538739] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "refresh_cache-eac4b8f9-9cd5-44a3-a12e-9ec22767b907" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.538739] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 621.619443] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "refresh_cache-abb9dc0e-ce18-49c8-a472-d39a3fcec887" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.619443] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Instance network_info: |[{"id": "d06bc826-4580-47a4-a3b4-841380ee155f", "address": "fa:16:3e:37:92:b6", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd06bc826-45", "ovs_interfaceid": "d06bc826-4580-47a4-a3b4-841380ee155f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 621.620070] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:92:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c13fd8bc-e797-42fe-94ed-6370d3467a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd06bc826-4580-47a4-a3b4-841380ee155f', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 621.630562] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Creating folder: Project (24d8b15ee5034b33a4c741412bcc71e0). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 621.632877] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e072b783-5a94-4aec-b486-ddad642d61af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.648042] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Created folder: Project (24d8b15ee5034b33a4c741412bcc71e0) in parent group-v992574. [ 621.648042] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Creating folder: Instances. Parent ref: group-v992601. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 621.648279] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07fb5c90-82c5-4f45-a033-d0a1ce313063 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.658881] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Created folder: Instances in parent group-v992601. [ 621.659156] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 621.659362] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 621.659583] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99a06d13-fd0c-489c-b0e4-cd76a28ace9d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.687467] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 621.687467] env[65107]: value = "task-5102219" [ 621.687467] env[65107]: _type = "Task" [ 621.687467] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.697164] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102219, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.764532] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102215, 'name': CreateVM_Task, 'duration_secs': 0.64248} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.764532] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 621.765170] env[65107]: WARNING neutronclient.v2_0.client [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.765673] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.765878] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.766287] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 621.766687] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ef1d3b9-8917-429b-a8b5-441eefe0caef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.776604] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 621.776604] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52421976-b92e-940e-9648-761529b22630" [ 621.776604] env[65107]: _type = "Task" [ 621.776604] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.780419] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127915} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.785251] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.787061] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ca00c4-8dd5-492a-8124-1d29eec237c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.817944] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] 8306ff90-b8bd-4270-8133-96abe483156b/8306ff90-b8bd-4270-8133-96abe483156b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.828467] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e97b7e3-5386-4a95-a195-f345f062ccba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.846666] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52421976-b92e-940e-9648-761529b22630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.854439] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 621.854439] env[65107]: value = "task-5102220" [ 621.854439] env[65107]: _type = "Task" [ 621.854439] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.870588] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102220, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.888998] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 621.889606] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 621.893142] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.900s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 621.894319] env[65107]: INFO nova.compute.claims [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.897411] env[65107]: DEBUG oslo_concurrency.lockutils [req-7d701df2-c428-490b-a33b-fc5e44b8ba8b req-88e1b0fe-9490-45d5-8fbb-954789719f75 service nova] Releasing lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.022208] env[65107]: WARNING neutronclient.v2_0.client [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.022935] env[65107]: WARNING openstack [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.024985] env[65107]: WARNING openstack [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.045939] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.045939] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.078292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.110924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.111185] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.124156] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 622.148926] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.149356] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.198995] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102219, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.201304] env[65107]: WARNING openstack [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.201715] env[65107]: WARNING openstack [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.245366] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.246131] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.246366] env[65107]: WARNING openstack [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.279403] env[65107]: WARNING neutronclient.v2_0.client [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.280166] env[65107]: WARNING openstack [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.280561] env[65107]: WARNING openstack [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.303665] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52421976-b92e-940e-9648-761529b22630, 'name': SearchDatastore_Task, 'duration_secs': 0.020653} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.304146] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.304484] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 622.304736] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.304991] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 622.305305] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 622.305618] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf40dc1d-846d-45ff-9c9f-a0646dd8b726 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.317342] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 622.317463] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 622.318391] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a8e84f5-02f9-4605-8491-9b56b535f09c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.325759] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 622.325759] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c2061a-ebb2-88b4-9f32-b6a5e76051a8" [ 622.325759] env[65107]: _type = "Task" [ 622.325759] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.340071] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c2061a-ebb2-88b4-9f32-b6a5e76051a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.368319] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.401117] env[65107]: DEBUG nova.compute.utils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 622.404158] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 622.404379] env[65107]: DEBUG nova.network.neutron [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 622.404710] env[65107]: WARNING neutronclient.v2_0.client [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.405037] env[65107]: WARNING neutronclient.v2_0.client [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.405641] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.405967] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.432889] env[65107]: DEBUG nova.compute.manager [req-2fb11938-243d-48be-acf9-237ed84fc7b5 req-b51bd952-e411-472d-9ae4-d147c899288e service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Received event network-vif-plugged-d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 622.433304] env[65107]: DEBUG oslo_concurrency.lockutils [req-2fb11938-243d-48be-acf9-237ed84fc7b5 req-b51bd952-e411-472d-9ae4-d147c899288e service nova] Acquiring lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.434047] env[65107]: DEBUG oslo_concurrency.lockutils [req-2fb11938-243d-48be-acf9-237ed84fc7b5 req-b51bd952-e411-472d-9ae4-d147c899288e service nova] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.434047] env[65107]: DEBUG oslo_concurrency.lockutils [req-2fb11938-243d-48be-acf9-237ed84fc7b5 req-b51bd952-e411-472d-9ae4-d147c899288e service nova] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.434047] env[65107]: DEBUG nova.compute.manager [req-2fb11938-243d-48be-acf9-237ed84fc7b5 req-b51bd952-e411-472d-9ae4-d147c899288e service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] No waiting events found dispatching network-vif-plugged-d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 622.434240] env[65107]: WARNING nova.compute.manager [req-2fb11938-243d-48be-acf9-237ed84fc7b5 req-b51bd952-e411-472d-9ae4-d147c899288e service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Received unexpected event network-vif-plugged-d06bc826-4580-47a4-a3b4-841380ee155f for instance with vm_state building and task_state spawning. [ 622.483603] env[65107]: DEBUG nova.network.neutron [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Updating instance_info_cache with network_info: [{"id": "b41e2bb0-3a13-473d-ac19-0e593444f482", "address": "fa:16:3e:42:de:10", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb41e2bb0-3a", "ovs_interfaceid": "b41e2bb0-3a13-473d-ac19-0e593444f482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 622.521507] env[65107]: DEBUG nova.network.neutron [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updated VIF entry in instance network info cache for port e41be772-a45d-410b-acaf-09d30f9afa9d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 622.522032] env[65107]: DEBUG nova.network.neutron [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updating instance_info_cache with network_info: [{"id": "e41be772-a45d-410b-acaf-09d30f9afa9d", "address": "fa:16:3e:f7:84:8e", "network": {"id": "9a766a32-cd65-4a5e-b338-308b5e640847", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-2025951181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d98fae2faa0a40c798e251737ccb31c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41be772-a4", "ovs_interfaceid": "e41be772-a45d-410b-acaf-09d30f9afa9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 622.615117] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 622.655145] env[65107]: DEBUG nova.policy [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee5cd4f990b8494eba6e0d07a99a9949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5613159bd662493aa8bf523e0d8af6ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 622.701177] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102219, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.839063] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c2061a-ebb2-88b4-9f32-b6a5e76051a8, 'name': SearchDatastore_Task, 'duration_secs': 0.041196} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.840289] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fce137ab-aa31-40a0-9ec7-f3077c9c8040 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.846805] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 622.846805] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d37c0-6d08-6962-1c9d-35e9ff3527fb" [ 622.846805] env[65107]: _type = "Task" [ 622.846805] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.856372] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d37c0-6d08-6962-1c9d-35e9ff3527fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.865925] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102220, 'name': ReconfigVM_Task, 'duration_secs': 0.64664} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.866309] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Reconfigured VM instance instance-00000008 to attach disk [datastore1] 8306ff90-b8bd-4270-8133-96abe483156b/8306ff90-b8bd-4270-8133-96abe483156b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.866996] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c2118662-7e60-4e2a-b931-9e1a23564777 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.875607] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 622.875607] env[65107]: value = "task-5102221" [ 622.875607] env[65107]: _type = "Task" [ 622.875607] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.885839] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102221, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.915743] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 622.988698] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "refresh_cache-eac4b8f9-9cd5-44a3-a12e-9ec22767b907" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.988807] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Instance network_info: |[{"id": "b41e2bb0-3a13-473d-ac19-0e593444f482", "address": "fa:16:3e:42:de:10", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb41e2bb0-3a", "ovs_interfaceid": "b41e2bb0-3a13-473d-ac19-0e593444f482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 622.990486] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:de:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c13fd8bc-e797-42fe-94ed-6370d3467a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b41e2bb0-3a13-473d-ac19-0e593444f482', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 622.998141] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 622.999254] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.999873] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7fda88f9-194d-437d-bd7e-ca7b0a15d046 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.025138] env[65107]: DEBUG oslo_concurrency.lockutils [req-8ca4ce49-ab79-4bae-93bc-5e06e244e888 req-9e1c8bfe-b98e-4400-a5cd-1846a65cc5ca service nova] Releasing lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.028724] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 623.028724] env[65107]: value = "task-5102222" [ 623.028724] env[65107]: _type = "Task" [ 623.028724] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.040373] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102222, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.150025] env[65107]: DEBUG nova.network.neutron [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Successfully created port: 20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 623.153619] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.202786] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102219, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.259281] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15af8c4d-f31f-4468-9e76-f8287fd5ffff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.270545] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7924ec35-4bbf-43bb-a54e-e1562377381e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.304792] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236fd73a-cd7d-40b3-85dd-af0604448f58 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.312807] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1368bfa-da24-45d7-8cf6-01c232c12bdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.331389] env[65107]: DEBUG nova.compute.provider_tree [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.362438] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d37c0-6d08-6962-1c9d-35e9ff3527fb, 'name': SearchDatastore_Task, 'duration_secs': 0.011913} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.362438] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.362438] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ccd68268-d0fc-406e-896f-c61b114dd75d/ccd68268-d0fc-406e-896f-c61b114dd75d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 623.363152] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6620f7e4-6eb3-4859-9841-00189465687a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.373538] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 623.373538] env[65107]: value = "task-5102227" [ 623.373538] env[65107]: _type = "Task" [ 623.373538] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.388577] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.392358] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102221, 'name': Rename_Task, 'duration_secs': 0.248688} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.392715] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 623.393051] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e47fa90-bbc3-40bd-bec6-dfe4209f405a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.400842] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 623.400842] env[65107]: value = "task-5102228" [ 623.400842] env[65107]: _type = "Task" [ 623.400842] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.412568] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.541685] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102222, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.679812] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Acquiring lock "31195587-86ce-4379-946d-1abb51c35f08" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.680357] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lock "31195587-86ce-4379-946d-1abb51c35f08" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.680625] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Acquiring lock "31195587-86ce-4379-946d-1abb51c35f08-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.680817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lock "31195587-86ce-4379-946d-1abb51c35f08-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.681013] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lock "31195587-86ce-4379-946d-1abb51c35f08-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.683609] env[65107]: INFO nova.compute.manager [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Terminating instance [ 623.702544] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102219, 'name': CreateVM_Task, 'duration_secs': 1.573875} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.703019] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 623.703697] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 623.703898] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.704061] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.704369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 623.704950] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a410fbb-505a-4457-8dc6-727c3969aa2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.711237] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 623.711237] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5250867c-7d95-dff9-7006-021ac3349146" [ 623.711237] env[65107]: _type = "Task" [ 623.711237] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.721112] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5250867c-7d95-dff9-7006-021ac3349146, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.839985] env[65107]: DEBUG nova.scheduler.client.report [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 623.885815] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.912903] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.932528] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 623.962279] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 623.962554] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.962709] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 623.962950] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.963324] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 623.963324] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 623.963612] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.963747] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 623.963942] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 623.964159] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 623.964345] env[65107]: DEBUG nova.virt.hardware [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 623.965373] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cab5235-4a4c-4793-a19a-a518216ceec3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.974760] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a7c0cd-b698-44d8-8e7e-db6d5cced4fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.040041] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102222, 'name': CreateVM_Task, 'duration_secs': 0.556708} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.040205] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 624.040708] env[65107]: WARNING neutronclient.v2_0.client [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 624.041064] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.188889] env[65107]: DEBUG nova.compute.manager [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 624.189163] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.190069] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2a282f-9c41-4742-a501-3a0ce36691e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.200894] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 624.201404] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f891181-3e62-46a6-ae90-f6988f8da885 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.210397] env[65107]: DEBUG oslo_vmware.api [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Waiting for the task: (returnval){ [ 624.210397] env[65107]: value = "task-5102229" [ 624.210397] env[65107]: _type = "Task" [ 624.210397] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.228851] env[65107]: DEBUG oslo_vmware.api [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Task: {'id': task-5102229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.234682] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5250867c-7d95-dff9-7006-021ac3349146, 'name': SearchDatastore_Task, 'duration_secs': 0.015013} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.235306] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.235882] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.236294] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.236583] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.236867] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.237380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.237766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 624.238187] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a2031b2-d08c-440a-8f6b-0095f8cbcbaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.241393] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9637387-8a8a-424c-9f30-c9346cad08bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.250100] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 624.250100] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a85eb-ab37-102a-0d03-fd40dd05d197" [ 624.250100] env[65107]: _type = "Task" [ 624.250100] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.258806] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.261289] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 624.263992] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14c57763-f37e-46e6-aae9-f074e57665cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.267270] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a85eb-ab37-102a-0d03-fd40dd05d197, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.271650] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 624.271650] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5245c28e-94d2-b7e3-a282-1b8c3ddee4bc" [ 624.271650] env[65107]: _type = "Task" [ 624.271650] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.282009] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5245c28e-94d2-b7e3-a282-1b8c3ddee4bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.343109] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.343919] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 624.348377] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.270s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.352023] env[65107]: INFO nova.compute.claims [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.389713] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102227, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.419172] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.720913] env[65107]: DEBUG oslo_vmware.api [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Task: {'id': task-5102229, 'name': PowerOffVM_Task, 'duration_secs': 0.35879} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.721233] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 624.721377] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 624.721637] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d87bef8-9391-4cee-8f47-8896f2ba2806 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.763527] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a85eb-ab37-102a-0d03-fd40dd05d197, 'name': SearchDatastore_Task, 'duration_secs': 0.021223} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.763878] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.764201] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.764426] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.783232] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5245c28e-94d2-b7e3-a282-1b8c3ddee4bc, 'name': SearchDatastore_Task, 'duration_secs': 0.017296} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.784222] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe76fc37-d06d-4289-abf9-805d688848bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.791649] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 624.791649] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c8fd38-3300-bb88-785b-8315264a2604" [ 624.791649] env[65107]: _type = "Task" [ 624.791649] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.805511] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c8fd38-3300-bb88-785b-8315264a2604, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.841813] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 624.842140] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 624.842783] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Deleting the datastore file [datastore2] 31195587-86ce-4379-946d-1abb51c35f08 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 624.843191] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89289084-8ca1-4dd4-907f-43ee3c3c0098 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.852338] env[65107]: DEBUG oslo_vmware.api [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Waiting for the task: (returnval){ [ 624.852338] env[65107]: value = "task-5102231" [ 624.852338] env[65107]: _type = "Task" [ 624.852338] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.858530] env[65107]: DEBUG nova.compute.utils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 624.862562] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 624.862562] env[65107]: DEBUG nova.network.neutron [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 624.863040] env[65107]: WARNING neutronclient.v2_0.client [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 624.864768] env[65107]: WARNING neutronclient.v2_0.client [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 624.864898] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.865311] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.892107] env[65107]: DEBUG oslo_vmware.api [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Task: {'id': task-5102231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.899430] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102227, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.303022} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.899725] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ccd68268-d0fc-406e-896f-c61b114dd75d/ccd68268-d0fc-406e-896f-c61b114dd75d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 624.899992] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 624.900289] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb2f5258-930c-4bfb-8b2a-d921235cd332 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.909042] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 624.909042] env[65107]: value = "task-5102232" [ 624.909042] env[65107]: _type = "Task" [ 624.909042] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.915543] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102228, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.927429] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.079273] env[65107]: DEBUG nova.network.neutron [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Successfully updated port: 20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 625.153783] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e785db4-0132-4ae0-a552-debafb3f3bc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.163740] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af10b1d1-631f-46d3-81ea-7f8f60d2975d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.200800] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5af48f-5fc0-47db-bb0b-0f2ba006901c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.208917] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8fcf46-541c-470c-86d7-600ca5fd280d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.224069] env[65107]: DEBUG nova.compute.provider_tree [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.304151] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c8fd38-3300-bb88-785b-8315264a2604, 'name': SearchDatastore_Task, 'duration_secs': 0.021449} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.304910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.304910] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] abb9dc0e-ce18-49c8-a472-d39a3fcec887/abb9dc0e-ce18-49c8-a472-d39a3fcec887.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.305167] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.305167] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.305592] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-728a9e2e-5af3-482e-8d2d-7c5061bb9046 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.307534] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ede2a4f-f603-4ce7-a7ae-c32a9015703f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.316335] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 625.316335] env[65107]: value = "task-5102233" [ 625.316335] env[65107]: _type = "Task" [ 625.316335] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.321113] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.321430] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.322642] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-834066b9-0ec0-4ec7-b031-409cf6fef681 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.328489] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.332696] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 625.332696] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5251af9e-6122-6776-b3b3-b64d9f9adaba" [ 625.332696] env[65107]: _type = "Task" [ 625.332696] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.341911] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5251af9e-6122-6776-b3b3-b64d9f9adaba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.363573] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 625.366928] env[65107]: DEBUG oslo_vmware.api [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Task: {'id': task-5102231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34766} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.367612] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.367829] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 625.368056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.369225] env[65107]: INFO nova.compute.manager [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Took 1.18 seconds to destroy the instance on the hypervisor. [ 625.369225] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 625.369225] env[65107]: DEBUG nova.compute.manager [-] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 625.369225] env[65107]: DEBUG nova.network.neutron [-] [instance: 31195587-86ce-4379-946d-1abb51c35f08] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 625.369374] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.370023] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.370291] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.415866] env[65107]: DEBUG oslo_vmware.api [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102228, 'name': PowerOnVM_Task, 'duration_secs': 1.737379} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.415866] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 625.415866] env[65107]: INFO nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Took 12.88 seconds to spawn the instance on the hypervisor. [ 625.416118] env[65107]: DEBUG nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 625.420131] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd5fc81-7dbb-4d7c-86db-e4a00f6efa9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.429784] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06922} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.431525] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 625.437993] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acd5065-759c-4348-8ea3-13a2dcb94156 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.463178] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] ccd68268-d0fc-406e-896f-c61b114dd75d/ccd68268-d0fc-406e-896f-c61b114dd75d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 625.463650] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a4339fb-83b7-4490-a29c-168d8484716e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.487221] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 625.487221] env[65107]: value = "task-5102234" [ 625.487221] env[65107]: _type = "Task" [ 625.487221] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.500725] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102234, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.583956] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "refresh_cache-7a3009bf-54a2-4565-a1aa-1d19286a4810" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.584208] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquired lock "refresh_cache-7a3009bf-54a2-4565-a1aa-1d19286a4810" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.584403] env[65107]: DEBUG nova.network.neutron [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 625.728887] env[65107]: DEBUG nova.scheduler.client.report [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 625.750600] env[65107]: DEBUG nova.policy [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e7e78dbbfc041da983b34a950e383b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '634f89d91e1443c7a0eb1314ecb17232', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 625.792016] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.832658] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102233, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.853258] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5251af9e-6122-6776-b3b3-b64d9f9adaba, 'name': SearchDatastore_Task, 'duration_secs': 0.021783} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.854458] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e3991ff-951e-49e4-bc88-d17bf605a9e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.865553] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 625.865553] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5293e0cd-8c20-6b8e-6cd6-260058bcc5cd" [ 625.865553] env[65107]: _type = "Task" [ 625.865553] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.890812] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5293e0cd-8c20-6b8e-6cd6-260058bcc5cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.956712] env[65107]: INFO nova.compute.manager [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Took 25.44 seconds to build instance. [ 626.008637] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102234, 'name': ReconfigVM_Task, 'duration_secs': 0.355201} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.009241] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Reconfigured VM instance instance-00000009 to attach disk [datastore1] ccd68268-d0fc-406e-896f-c61b114dd75d/ccd68268-d0fc-406e-896f-c61b114dd75d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 626.010477] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d4ae6a5-4238-4ab6-9758-dd1752bd6c23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.025342] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 626.025342] env[65107]: value = "task-5102235" [ 626.025342] env[65107]: _type = "Task" [ 626.025342] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.037635] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102235, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.090238] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.091507] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.235744] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.887s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.236669] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 626.239910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.086s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.241565] env[65107]: INFO nova.compute.claims [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.329363] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.751652} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.329605] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] abb9dc0e-ce18-49c8-a472-d39a3fcec887/abb9dc0e-ce18-49c8-a472-d39a3fcec887.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 626.329724] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 626.330185] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-383aeabb-6164-4757-a241-c8900756f355 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.338935] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 626.338935] env[65107]: value = "task-5102236" [ 626.338935] env[65107]: _type = "Task" [ 626.338935] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.349660] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102236, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.381284] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 626.385565] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5293e0cd-8c20-6b8e-6cd6-260058bcc5cd, 'name': SearchDatastore_Task, 'duration_secs': 0.058313} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.386468] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.386731] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] eac4b8f9-9cd5-44a3-a12e-9ec22767b907/eac4b8f9-9cd5-44a3-a12e-9ec22767b907.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 626.387040] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ce60763-f497-4cb4-b090-1061d88f316c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.391871] env[65107]: INFO nova.compute.manager [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Rebuilding instance [ 626.401328] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 626.401328] env[65107]: value = "task-5102238" [ 626.401328] env[65107]: _type = "Task" [ 626.401328] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.414432] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.416961] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:21:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1896394497',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-952575834',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 626.417203] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 626.417357] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 626.417646] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 626.417728] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 626.417852] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 626.418101] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.418273] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 626.418440] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 626.418600] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 626.418768] env[65107]: DEBUG nova.virt.hardware [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 626.420838] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86da4ceb-9fd2-4fa8-8e4f-529576b5fa92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.430877] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2902d9ae-5c4d-4282-bcfb-fdbeb59a853c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.461189] env[65107]: DEBUG oslo_concurrency.lockutils [None req-54b14cf9-3d6e-468d-8f25-782fb211f9cf tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "8306ff90-b8bd-4270-8133-96abe483156b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.963s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.467432] env[65107]: DEBUG nova.compute.manager [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 626.468666] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae759c7-a8a0-407c-badf-e304de818ed8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.538427] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102235, 'name': Rename_Task, 'duration_secs': 0.171808} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.538892] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 626.540271] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c6e82dc-4e5f-428e-88c7-b727d1166a66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.549399] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 626.549399] env[65107]: value = "task-5102239" [ 626.549399] env[65107]: _type = "Task" [ 626.549399] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.562388] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102239, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.748294] env[65107]: DEBUG nova.compute.utils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 626.753599] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 626.754090] env[65107]: DEBUG nova.network.neutron [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 626.755292] env[65107]: WARNING neutronclient.v2_0.client [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.756986] env[65107]: WARNING neutronclient.v2_0.client [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.756986] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.756986] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.857834] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102236, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077849} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.858069] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.858999] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ed4c2c-f274-4470-a057-23a8a5c4831d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.891136] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] abb9dc0e-ce18-49c8-a472-d39a3fcec887/abb9dc0e-ce18-49c8-a472-d39a3fcec887.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.891409] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bc5f3ef-0f6a-4fc9-80d2-136b0e923998 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.920803] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 626.920803] env[65107]: value = "task-5102240" [ 626.920803] env[65107]: _type = "Task" [ 626.920803] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.931740] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102238, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.948295] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102240, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.065076] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102239, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.160749] env[65107]: DEBUG nova.network.neutron [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 627.255184] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 627.425214] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102238, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.804821} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.425474] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] eac4b8f9-9cd5-44a3-a12e-9ec22767b907/eac4b8f9-9cd5-44a3-a12e-9ec22767b907.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 627.425781] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.425885] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3573183c-6e31-4b55-ae89-aa75e5f25750 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.436354] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 627.436354] env[65107]: value = "task-5102241" [ 627.436354] env[65107]: _type = "Task" [ 627.436354] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.439873] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102240, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.448654] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.488685] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 627.489093] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bbdf1aa-0541-4562-9479-d841902ad58f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.497230] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 627.497230] env[65107]: value = "task-5102242" [ 627.497230] env[65107]: _type = "Task" [ 627.497230] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.510551] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.544906] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35a223b-b7e6-4ea5-844a-c2921fa6b218 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.558652] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c8da3b-09d2-4bb6-9675-905206929b57 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.569175] env[65107]: DEBUG oslo_vmware.api [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102239, 'name': PowerOnVM_Task, 'duration_secs': 0.635877} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.575616] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 627.575616] env[65107]: INFO nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Took 12.66 seconds to spawn the instance on the hypervisor. [ 627.575616] env[65107]: DEBUG nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 627.575616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 627.576370] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.576481] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafc6d5e-dcb4-4d50-9cf4-26a2f8e05fdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.611938] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c7541a-f0f0-4997-88f7-ea3ea9b409cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.626553] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115543ec-531e-47a2-80d1-ab68b2449420 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.632050] env[65107]: DEBUG nova.network.neutron [-] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 627.646772] env[65107]: DEBUG nova.compute.provider_tree [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.684710] env[65107]: DEBUG oslo_concurrency.lockutils [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 627.684961] env[65107]: DEBUG oslo_concurrency.lockutils [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.685261] env[65107]: INFO nova.compute.manager [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Rebooting instance [ 627.948568] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102240, 'name': ReconfigVM_Task, 'duration_secs': 1.016039} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.949549] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Reconfigured VM instance instance-0000000a to attach disk [datastore1] abb9dc0e-ce18-49c8-a472-d39a3fcec887/abb9dc0e-ce18-49c8-a472-d39a3fcec887.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.950120] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-017e4f5a-3d52-43c5-bd7b-913ee627d1ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.955889] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069212} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.956264] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 627.957153] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adecb58a-ce57-4c96-bf93-87486376f3f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.962945] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 627.962945] env[65107]: value = "task-5102243" [ 627.962945] env[65107]: _type = "Task" [ 627.962945] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.986691] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] eac4b8f9-9cd5-44a3-a12e-9ec22767b907/eac4b8f9-9cd5-44a3-a12e-9ec22767b907.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 627.987108] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29dd1117-e1f0-4c18-88c4-5f618aec730c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.010321] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102243, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.020894] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102242, 'name': PowerOffVM_Task, 'duration_secs': 0.147902} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.020894] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 628.020894] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.021545] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 628.021545] env[65107]: value = "task-5102244" [ 628.021545] env[65107]: _type = "Task" [ 628.021545] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.022021] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c71f07-47c4-4d13-a600-0c03aa9fb03f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.035374] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102244, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.037857] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 628.038151] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85bbfdb9-a113-4605-99f4-7dbb01765dbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.066763] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 628.066983] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 628.067267] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Deleting the datastore file [datastore1] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 628.067584] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ba64067-7994-4349-bc8a-e0c006bd3dba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.076049] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 628.076049] env[65107]: value = "task-5102246" [ 628.076049] env[65107]: _type = "Task" [ 628.076049] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.086527] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.108702] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 628.131688] env[65107]: INFO nova.compute.manager [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Took 23.45 seconds to build instance. [ 628.137351] env[65107]: INFO nova.compute.manager [-] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Took 2.77 seconds to deallocate network for instance. [ 628.157798] env[65107]: DEBUG nova.scheduler.client.report [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 628.240603] env[65107]: DEBUG oslo_concurrency.lockutils [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.241230] env[65107]: DEBUG oslo_concurrency.lockutils [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquired lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.242364] env[65107]: DEBUG nova.network.neutron [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 628.263785] env[65107]: DEBUG nova.network.neutron [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Successfully created port: 3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 628.266140] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 628.326019] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 628.326353] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 628.326599] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 628.326682] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 628.327317] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 628.327317] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 628.327317] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.327317] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 628.327546] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 628.327756] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 628.327756] env[65107]: DEBUG nova.virt.hardware [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 628.329156] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8da42d-b967-4fa5-98b4-5f1b07193639 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.342048] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6664d4cf-5174-4852-a02f-42c2adacbead {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.443470] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.444250] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.475826] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102243, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.539583] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102244, 'name': ReconfigVM_Task, 'duration_secs': 0.363245} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.539583] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Reconfigured VM instance instance-0000000b to attach disk [datastore1] eac4b8f9-9cd5-44a3-a12e-9ec22767b907/eac4b8f9-9cd5-44a3-a12e-9ec22767b907.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 628.541844] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83807d37-39b2-43e4-b8d3-7c17546ebcee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.551750] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 628.551750] env[65107]: value = "task-5102248" [ 628.551750] env[65107]: _type = "Task" [ 628.551750] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.561603] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102248, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.590551] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203047} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.591067] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 628.591318] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 628.591509] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.634496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3babe0bb-30d6-4056-9f53-5a9200f0ed4e tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.970s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.642834] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 628.645913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 628.664956] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.664956] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 628.667578] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.025s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 628.669452] env[65107]: INFO nova.compute.claims [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.741768] env[65107]: DEBUG nova.policy [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9d79bf2d14e4c3d9eb6d38d6f38ba13', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8492fd3bcbf0485c84e4bf980f3d2809', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 628.745470] env[65107]: WARNING neutronclient.v2_0.client [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.746153] env[65107]: WARNING openstack [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.746501] env[65107]: WARNING openstack [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.974344] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102243, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.062324] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102248, 'name': Rename_Task, 'duration_secs': 0.167907} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.062607] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 629.062911] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e9a0fed-a234-45d9-9b1c-c782b32356a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.072028] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 629.072028] env[65107]: value = "task-5102249" [ 629.072028] env[65107]: _type = "Task" [ 629.072028] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.083534] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102249, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.177986] env[65107]: DEBUG nova.compute.utils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 629.179585] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 629.179787] env[65107]: DEBUG nova.network.neutron [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 629.180119] env[65107]: WARNING neutronclient.v2_0.client [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.180418] env[65107]: WARNING neutronclient.v2_0.client [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.182584] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.183053] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.475752] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102243, 'name': Rename_Task, 'duration_secs': 1.26733} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.476130] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 629.476324] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a30ebf7-cc74-48b4-955e-a84d498e6608 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.488358] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 629.488358] env[65107]: value = "task-5102250" [ 629.488358] env[65107]: _type = "Task" [ 629.488358] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.500526] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.584664] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102249, 'name': PowerOnVM_Task, 'duration_secs': 0.511117} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.585517] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 629.585517] env[65107]: INFO nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Took 9.91 seconds to spawn the instance on the hypervisor. [ 629.585722] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 629.587109] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c317b16c-36df-4d05-a80a-3435e2d46c30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.654764] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 629.655010] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.656443] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 629.656889] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.656889] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 629.657034] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 629.657286] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.657714] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 629.657933] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 629.658564] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 629.658564] env[65107]: DEBUG nova.virt.hardware [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 629.663503] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e53e133-0955-4702-935a-51867c0054d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.672995] env[65107]: DEBUG nova.compute.manager [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 629.683069] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a78355-4449-4ed4-9fcc-1762514dfe75 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.700969] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 629.703944] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 629.709819] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 629.713327] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 629.713900] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e4fc19a-ea4e-466a-8787-ddffc7493a46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.737254] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 629.737254] env[65107]: value = "task-5102251" [ 629.737254] env[65107]: _type = "Task" [ 629.737254] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.747208] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102251, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.005128] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102250, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.060875] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfd7664-ca8e-4850-8e5f-b6023fe562af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.071329] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d98c49-b239-4749-a0c4-3054092a990c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.124824] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7042a307-fd67-4a9e-9a6a-ab0c9175d501 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.132568] env[65107]: INFO nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Took 25.08 seconds to build instance. [ 630.136291] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03704b18-00e8-4f22-8542-c35d823a9ab0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.152461] env[65107]: DEBUG nova.compute.provider_tree [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.193446] env[65107]: DEBUG nova.network.neutron [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Successfully created port: 0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 630.200508] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.248862] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102251, 'name': CreateVM_Task, 'duration_secs': 0.335267} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.249065] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 630.249497] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.249752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.250396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 630.250584] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2cc4243-2458-4a21-b3d7-c994b9c94d70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.257243] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 630.257243] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5209e853-3fb3-e2f7-6f28-09059ca67887" [ 630.257243] env[65107]: _type = "Task" [ 630.257243] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.270696] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5209e853-3fb3-e2f7-6f28-09059ca67887, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.296968] env[65107]: DEBUG nova.policy [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ccab6477dd44878f78a9d6428d3cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4778fe9152224fd29d1f6220a19b5a36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 630.348597] env[65107]: WARNING neutronclient.v2_0.client [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.349282] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.349606] env[65107]: WARNING openstack [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.428345] env[65107]: DEBUG nova.network.neutron [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Successfully updated port: 3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 630.504897] env[65107]: DEBUG oslo_vmware.api [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102250, 'name': PowerOnVM_Task, 'duration_secs': 0.63142} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.505327] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 630.505835] env[65107]: INFO nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Took 13.18 seconds to spawn the instance on the hypervisor. [ 630.506142] env[65107]: DEBUG nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 630.507759] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59ee12a-d0a5-4ccf-924e-86bb807ed632 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.632542] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.597s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.660373] env[65107]: DEBUG nova.scheduler.client.report [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 630.717615] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 630.748093] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 630.748370] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 630.748370] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 630.748519] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 630.748659] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 630.748802] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 630.749223] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.749223] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 630.749349] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 630.749484] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 630.749646] env[65107]: DEBUG nova.virt.hardware [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 630.750722] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd5424a-b51a-45f8-9afd-1f797d37cd0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.763763] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2ca7ea-ace2-4029-9461-6f8d56aee7bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.774981] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5209e853-3fb3-e2f7-6f28-09059ca67887, 'name': SearchDatastore_Task, 'duration_secs': 0.015273} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.783374] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.783702] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 630.783899] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.784075] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.784259] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 630.784746] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22ba19ea-80bb-42c5-954f-34f5091c571d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.798437] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 630.798437] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 630.798437] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6595f49c-9639-431a-8ddd-162852453ccb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.807814] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 630.807814] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f770eb-58f4-ef44-d329-7842597d77a4" [ 630.807814] env[65107]: _type = "Task" [ 630.807814] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.819585] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f770eb-58f4-ef44-d329-7842597d77a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.871704] env[65107]: DEBUG nova.network.neutron [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Successfully created port: d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 630.934378] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.934571] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.934785] env[65107]: DEBUG nova.network.neutron [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 630.967609] env[65107]: WARNING openstack [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.968068] env[65107]: WARNING openstack [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.041525] env[65107]: INFO nova.compute.manager [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Took 26.03 seconds to build instance. [ 631.165915] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.166500] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 631.169022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.524s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.169297] env[65107]: DEBUG nova.objects.instance [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lazy-loading 'resources' on Instance uuid 31195587-86ce-4379-946d-1abb51c35f08 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 631.321033] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f770eb-58f4-ef44-d329-7842597d77a4, 'name': SearchDatastore_Task, 'duration_secs': 0.013304} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.321872] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c403c29-d54c-446c-a475-4cdc75c1c1e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.328732] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 631.328732] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52abdb30-5914-6e43-7ff2-f95141033c83" [ 631.328732] env[65107]: _type = "Task" [ 631.328732] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.337750] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52abdb30-5914-6e43-7ff2-f95141033c83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.438859] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.439436] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.544061] env[65107]: DEBUG oslo_concurrency.lockutils [None req-17cd3f6e-50fa-41fa-9360-31234ce986bc tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.543s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.674337] env[65107]: DEBUG nova.compute.utils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 631.679217] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 631.679739] env[65107]: DEBUG nova.network.neutron [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 631.679984] env[65107]: WARNING neutronclient.v2_0.client [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.680530] env[65107]: WARNING neutronclient.v2_0.client [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.680974] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.681346] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.844141] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52abdb30-5914-6e43-7ff2-f95141033c83, 'name': SearchDatastore_Task, 'duration_secs': 0.038623} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.845047] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.845047] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 631.845258] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc1fdfd7-9ab2-4bab-a33d-314a304be6a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.855894] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 631.855894] env[65107]: value = "task-5102253" [ 631.855894] env[65107]: _type = "Task" [ 631.855894] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.865805] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.922697] env[65107]: DEBUG nova.network.neutron [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Successfully updated port: 0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 631.961772] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1ddea6-b864-4970-93a1-3a037ab07542 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.971090] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d852eb-b4cb-4f45-a5d4-61786994d69f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.004143] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0738d86-2174-40b5-aa82-2aae5ad97f56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.014988] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f33f57b-15c2-4c47-8b38-15b9eb30338a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.035609] env[65107]: DEBUG nova.compute.provider_tree [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.178035] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 632.230369] env[65107]: DEBUG nova.network.neutron [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Updating instance_info_cache with network_info: [{"id": "20915f30-cfbb-4a09-bf64-495545e0096f", "address": "fa:16:3e:e6:94:75", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.54", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20915f30-cf", "ovs_interfaceid": "20915f30-cfbb-4a09-bf64-495545e0096f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.362832] env[65107]: DEBUG nova.network.neutron [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 632.369086] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508111} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.369488] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 632.369686] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.369984] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e37aa6cd-6aba-4075-8226-70c0529d2fdd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.378373] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 632.378373] env[65107]: value = "task-5102254" [ 632.378373] env[65107]: _type = "Task" [ 632.378373] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.388131] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102254, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.428068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "refresh_cache-f2814dd6-8be5-49b2-a030-f57e452c2e6d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.428068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired lock "refresh_cache-f2814dd6-8be5-49b2-a030-f57e452c2e6d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.428068] env[65107]: DEBUG nova.network.neutron [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 632.462075] env[65107]: WARNING neutronclient.v2_0.client [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.464056] env[65107]: WARNING openstack [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.464873] env[65107]: WARNING openstack [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.476158] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.476158] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.515361] env[65107]: DEBUG nova.policy [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9d79bf2d14e4c3d9eb6d38d6f38ba13', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8492fd3bcbf0485c84e4bf980f3d2809', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 632.524593] env[65107]: DEBUG nova.network.neutron [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Successfully updated port: d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 632.538891] env[65107]: DEBUG nova.scheduler.client.report [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.666192] env[65107]: DEBUG nova.network.neutron [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Updating instance_info_cache with network_info: [{"id": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "address": "fa:16:3e:ae:a6:0b", "network": {"id": "4f3c2189-942b-47c1-95bd-5064d80e3a76", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-78304541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4df06b84d1cd45b28b8ea33565ac45f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0d2101e-2d93-4310-a242-af2d9ecdaf9b", "external-id": "nsx-vlan-transportzone-121", "segmentation_id": 121, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2435e56-5f", "ovs_interfaceid": "d2435e56-5f0a-44e9-8154-2f2dd4103c55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.704687] env[65107]: WARNING neutronclient.v2_0.client [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.705330] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.706371] env[65107]: WARNING openstack [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.735147] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Releasing lock "refresh_cache-7a3009bf-54a2-4565-a1aa-1d19286a4810" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.736945] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Instance network_info: |[{"id": "20915f30-cfbb-4a09-bf64-495545e0096f", "address": "fa:16:3e:e6:94:75", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.54", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20915f30-cf", "ovs_interfaceid": "20915f30-cfbb-4a09-bf64-495545e0096f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 632.737464] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:94:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20915f30-cfbb-4a09-bf64-495545e0096f', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 632.747538] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Creating folder: Project (5613159bd662493aa8bf523e0d8af6ff). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.748410] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35f570b0-a7f2-4416-8410-bbd600e50839 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.762412] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Created folder: Project (5613159bd662493aa8bf523e0d8af6ff) in parent group-v992574. [ 632.762637] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Creating folder: Instances. Parent ref: group-v992610. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.762936] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b9db4ce-dd2f-4463-ae93-748c856bfcaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.773680] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Created folder: Instances in parent group-v992610. [ 632.774101] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 632.774353] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 632.774604] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecaf41a5-3bf5-4815-b81a-50e38556d845 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.794138] env[65107]: DEBUG nova.compute.manager [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Received event network-vif-plugged-02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 632.794138] env[65107]: DEBUG oslo_concurrency.lockutils [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Acquiring lock "ccd68268-d0fc-406e-896f-c61b114dd75d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.794138] env[65107]: DEBUG oslo_concurrency.lockutils [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.794138] env[65107]: DEBUG oslo_concurrency.lockutils [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.794138] env[65107]: DEBUG nova.compute.manager [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] No waiting events found dispatching network-vif-plugged-02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 632.794380] env[65107]: WARNING nova.compute.manager [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Received unexpected event network-vif-plugged-02108638-8e1e-447a-9b94-a45adfd32161 for instance with vm_state active and task_state None. [ 632.794380] env[65107]: DEBUG nova.compute.manager [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Received event network-changed-02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 632.794380] env[65107]: DEBUG nova.compute.manager [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Refreshing instance network info cache due to event network-changed-02108638-8e1e-447a-9b94-a45adfd32161. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 632.794732] env[65107]: DEBUG oslo_concurrency.lockutils [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Acquiring lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.794732] env[65107]: DEBUG oslo_concurrency.lockutils [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Acquired lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.795368] env[65107]: DEBUG nova.network.neutron [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Refreshing network info cache for port 02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 632.803322] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 632.803322] env[65107]: value = "task-5102258" [ 632.803322] env[65107]: _type = "Task" [ 632.803322] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.814573] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102258, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.896219] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102254, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07117} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.897502] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 632.898018] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fe66d4-40b7-4f57-afaa-e0baa17ab765 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.931640] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 632.933248] env[65107]: DEBUG nova.network.neutron [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Successfully created port: f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 632.938252] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.938586] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.945507] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05f0bf02-f267-44ed-ba1d-ae7c693cc729 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.964315] env[65107]: DEBUG nova.network.neutron [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Updating instance_info_cache with network_info: [{"id": "3bb17344-0031-4226-9117-1366ecf93330", "address": "fa:16:3e:06:8b:e4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bb17344-00", "ovs_interfaceid": "3bb17344-0031-4226-9117-1366ecf93330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.973446] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 632.973446] env[65107]: value = "task-5102259" [ 632.973446] env[65107]: _type = "Task" [ 632.973446] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.989723] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102259, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.025041] env[65107]: DEBUG nova.network.neutron [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 633.031267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-0691bdfe-ff23-471b-ad94-2a98adcecd4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.031267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-0691bdfe-ff23-471b-ad94-2a98adcecd4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.031679] env[65107]: DEBUG nova.network.neutron [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 633.047578] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.051799] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.851s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.078414] env[65107]: INFO nova.scheduler.client.report [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Deleted allocations for instance 31195587-86ce-4379-946d-1abb51c35f08 [ 633.094951] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.095417] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.116811] env[65107]: DEBUG nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Received event network-changed-d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 633.116811] env[65107]: DEBUG nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Refreshing instance network info cache due to event network-changed-d06bc826-4580-47a4-a3b4-841380ee155f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 633.116811] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Acquiring lock "refresh_cache-abb9dc0e-ce18-49c8-a472-d39a3fcec887" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.116811] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Acquired lock "refresh_cache-abb9dc0e-ce18-49c8-a472-d39a3fcec887" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.116811] env[65107]: DEBUG nova.network.neutron [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Refreshing network info cache for port d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 633.170892] env[65107]: DEBUG oslo_concurrency.lockutils [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Releasing lock "refresh_cache-4e70aaf3-5f1c-4a61-a790-a0a5e409170d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.188860] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 633.225569] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 633.225928] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 633.226142] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 633.226527] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 633.226639] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 633.226787] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 633.227071] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.227281] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 633.227493] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 633.227990] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 633.227990] env[65107]: DEBUG nova.virt.hardware [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 633.229061] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd9cddb-766d-4554-9e9e-762522ba7651 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.239616] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46555a0-e7d8-4587-88ac-c1676ff3e70c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.298568] env[65107]: WARNING neutronclient.v2_0.client [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.299317] env[65107]: WARNING openstack [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.299676] env[65107]: WARNING openstack [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.320346] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102258, 'name': CreateVM_Task, 'duration_secs': 0.438551} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.322108] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 633.324171] env[65107]: WARNING neutronclient.v2_0.client [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.324560] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.324797] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.325053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 633.325967] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "f58c3088-c821-4eec-be0a-81221debc98e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.326216] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "f58c3088-c821-4eec-be0a-81221debc98e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.328672] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ccbf3af-ea4a-49c6-91f8-f19858c884a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.335606] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 633.335606] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d0b86-2b19-30e3-3230-a90e250f02d8" [ 633.335606] env[65107]: _type = "Task" [ 633.335606] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.342974] env[65107]: WARNING neutronclient.v2_0.client [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.342974] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.342974] env[65107]: WARNING openstack [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.355542] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d0b86-2b19-30e3-3230-a90e250f02d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.466723] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Releasing lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.467142] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Instance network_info: |[{"id": "3bb17344-0031-4226-9117-1366ecf93330", "address": "fa:16:3e:06:8b:e4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bb17344-00", "ovs_interfaceid": "3bb17344-0031-4226-9117-1366ecf93330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 633.467613] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:8b:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bb17344-0031-4226-9117-1366ecf93330', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 633.475899] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Creating folder: Project (634f89d91e1443c7a0eb1314ecb17232). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 633.476373] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6259b877-7e8b-4906-ade7-f3b5b6a9e943 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.490478] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102259, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.497319] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Created folder: Project (634f89d91e1443c7a0eb1314ecb17232) in parent group-v992574. [ 633.497319] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Creating folder: Instances. Parent ref: group-v992613. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 633.500031] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73e3fa32-96ce-47d7-8c33-9948ae834f31 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.512456] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Created folder: Instances in parent group-v992613. [ 633.513149] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 633.513149] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 633.513366] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5611bfe4-8a12-4c9c-8a36-261db72ba31f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.535713] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.535976] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.544890] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 633.544890] env[65107]: value = "task-5102262" [ 633.544890] env[65107]: _type = "Task" [ 633.544890] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.561420] env[65107]: INFO nova.compute.claims [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.565689] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102262, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.592639] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98638bf8-0071-486d-acc5-67041a180688 tempest-DeleteServersAdminTestJSON-703408665 tempest-DeleteServersAdminTestJSON-703408665-project-admin] Lock "31195587-86ce-4379-946d-1abb51c35f08" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.912s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.596861] env[65107]: DEBUG nova.network.neutron [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Updating instance_info_cache with network_info: [{"id": "0b52bca6-ff8c-4322-9b18-1c1129186df6", "address": "fa:16:3e:64:fc:90", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b52bca6-ff", "ovs_interfaceid": "0b52bca6-ff8c-4322-9b18-1c1129186df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 633.612508] env[65107]: WARNING openstack [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.612777] env[65107]: WARNING openstack [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.620221] env[65107]: WARNING neutronclient.v2_0.client [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.620820] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.621270] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.678108] env[65107]: DEBUG nova.compute.manager [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 633.679370] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10987ea8-8b51-4f09-9b17-c3f947134530 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.745922] env[65107]: DEBUG nova.network.neutron [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 633.830577] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 633.847672] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d0b86-2b19-30e3-3230-a90e250f02d8, 'name': SearchDatastore_Task, 'duration_secs': 0.017922} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.855895] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.855895] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 633.855895] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.855895] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.856269] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 633.856269] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dadfd0e8-0865-4301-aa81-f4a437d8ef18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.861210] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 633.861358] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 633.862230] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-066acda0-b99f-4d88-9c6f-003261513cab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.869682] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 633.869682] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524f13b1-be90-9532-d8c7-78266a7982d5" [ 633.869682] env[65107]: _type = "Task" [ 633.869682] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.874452] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "0ef89aea-5373-4d07-9587-37890c08d35b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.874537] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "0ef89aea-5373-4d07-9587-37890c08d35b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.874697] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "0ef89aea-5373-4d07-9587-37890c08d35b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.874935] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "0ef89aea-5373-4d07-9587-37890c08d35b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.875057] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "0ef89aea-5373-4d07-9587-37890c08d35b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.878513] env[65107]: INFO nova.compute.manager [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Terminating instance [ 633.884188] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524f13b1-be90-9532-d8c7-78266a7982d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.903146] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.903533] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.947898] env[65107]: WARNING neutronclient.v2_0.client [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.948684] env[65107]: WARNING openstack [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.949121] env[65107]: WARNING openstack [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.989115] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102259, 'name': ReconfigVM_Task, 'duration_secs': 0.531762} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.989402] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Reconfigured VM instance instance-00000001 to attach disk [datastore2] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1/ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.990082] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9e3b3d4-55f5-483d-be4c-12e2baae6ad4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.998378] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 633.998378] env[65107]: value = "task-5102263" [ 633.998378] env[65107]: _type = "Task" [ 633.998378] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.012557] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102263, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.055479] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102262, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.068653] env[65107]: INFO nova.compute.resource_tracker [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating resource usage from migration 6662c375-2493-41ad-b464-55f88c6fa7c1 [ 634.100231] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Releasing lock "refresh_cache-f2814dd6-8be5-49b2-a030-f57e452c2e6d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.100604] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Instance network_info: |[{"id": "0b52bca6-ff8c-4322-9b18-1c1129186df6", "address": "fa:16:3e:64:fc:90", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b52bca6-ff", "ovs_interfaceid": "0b52bca6-ff8c-4322-9b18-1c1129186df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 634.103965] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:fc:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b52bca6-ff8c-4322-9b18-1c1129186df6', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.112910] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating folder: Project (8492fd3bcbf0485c84e4bf980f3d2809). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 634.113483] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86d033c2-9522-45d1-b3c3-552f2bbc4f25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.126277] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Created folder: Project (8492fd3bcbf0485c84e4bf980f3d2809) in parent group-v992574. [ 634.126471] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating folder: Instances. Parent ref: group-v992616. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 634.126776] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91c539b0-8032-4c28-b4ba-b7657cd3a37c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.146179] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.146482] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.155223] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Created folder: Instances in parent group-v992616. [ 634.155500] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 634.158149] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 634.159066] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7112d311-50a1-49a0-a51d-2ff59ca07736 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.184599] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 634.184599] env[65107]: value = "task-5102266" [ 634.184599] env[65107]: _type = "Task" [ 634.184599] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.195438] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102266, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.206336] env[65107]: WARNING neutronclient.v2_0.client [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.207861] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.207861] env[65107]: WARNING openstack [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.243560] env[65107]: DEBUG nova.network.neutron [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updated VIF entry in instance network info cache for port 02108638-8e1e-447a-9b94-a45adfd32161. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 634.244631] env[65107]: DEBUG nova.network.neutron [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updating instance_info_cache with network_info: [{"id": "02108638-8e1e-447a-9b94-a45adfd32161", "address": "fa:16:3e:49:92:f3", "network": {"id": "f011d711-ac42-4e0b-bcd5-8f569242741c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-475912691-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac04479a6164cf7b18b9ad304a0c9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02108638-8e", "ovs_interfaceid": "02108638-8e1e-447a-9b94-a45adfd32161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 634.353954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.386715] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524f13b1-be90-9532-d8c7-78266a7982d5, 'name': SearchDatastore_Task, 'duration_secs': 0.015223} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.388087] env[65107]: DEBUG nova.compute.manager [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 634.388302] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.388552] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-121faddb-0e0d-4dfb-8c51-624240389f84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.391446] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fe2309-b1f1-4582-943d-e38376ac852e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.399555] env[65107]: WARNING neutronclient.v2_0.client [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.400225] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.400612] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.413444] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 634.413753] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 634.413753] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52753890-7bd2-599b-7f32-d5972d78c4e5" [ 634.413753] env[65107]: _type = "Task" [ 634.413753] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.414514] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba95b03-fcd6-454e-9ba6-ba1dc00935c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.417146] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d184add4-bd98-425d-9c4d-6eb246326029 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.431968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b3b232-f902-4252-a975-bfe43c3cdd42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.436297] env[65107]: DEBUG oslo_vmware.api [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 634.436297] env[65107]: value = "task-5102267" [ 634.436297] env[65107]: _type = "Task" [ 634.436297] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.436517] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52753890-7bd2-599b-7f32-d5972d78c4e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.469753] env[65107]: DEBUG nova.network.neutron [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Updating instance_info_cache with network_info: [{"id": "d9b6f78e-710c-4b4c-a9bd-04632f29764a", "address": "fa:16:3e:0c:fb:ab", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9b6f78e-71", "ovs_interfaceid": "d9b6f78e-710c-4b4c-a9bd-04632f29764a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 634.472111] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a967593-83f4-430e-a5eb-af28841f9ba4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.478392] env[65107]: DEBUG oslo_vmware.api [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102267, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.484386] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873fc828-8fc7-410f-8269-4479c83df8a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.504366] env[65107]: DEBUG nova.compute.provider_tree [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.515900] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102263, 'name': Rename_Task, 'duration_secs': 0.418837} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.516912] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 634.517202] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-813c6be8-7dd2-45ea-a70e-f3d64400c776 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.524957] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Waiting for the task: (returnval){ [ 634.524957] env[65107]: value = "task-5102268" [ 634.524957] env[65107]: _type = "Task" [ 634.524957] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.538926] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.556008] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102262, 'name': CreateVM_Task, 'duration_secs': 0.67373} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.556287] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 634.556733] env[65107]: WARNING neutronclient.v2_0.client [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.557222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.557386] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.557708] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 634.557991] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-897b0887-141c-4dfe-92e3-1592a868b467 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.563780] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 634.563780] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5219bc5c-97aa-d504-6b90-ba5e3d85c143" [ 634.563780] env[65107]: _type = "Task" [ 634.563780] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.574157] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5219bc5c-97aa-d504-6b90-ba5e3d85c143, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.589688] env[65107]: DEBUG nova.network.neutron [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Updated VIF entry in instance network info cache for port d06bc826-4580-47a4-a3b4-841380ee155f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 634.590055] env[65107]: DEBUG nova.network.neutron [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Updating instance_info_cache with network_info: [{"id": "d06bc826-4580-47a4-a3b4-841380ee155f", "address": "fa:16:3e:37:92:b6", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd06bc826-45", "ovs_interfaceid": "d06bc826-4580-47a4-a3b4-841380ee155f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 634.658628] env[65107]: DEBUG nova.network.neutron [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Successfully updated port: f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 634.699042] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102266, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.699042] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795ac6f8-34ed-4f0d-ac12-4e4246586500 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.708919] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Doing hard reboot of VM {{(pid=65107) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 634.709295] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-1fa1db5a-6a7d-4e66-a050-01792afa4273 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.718646] env[65107]: DEBUG oslo_vmware.api [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 634.718646] env[65107]: value = "task-5102270" [ 634.718646] env[65107]: _type = "Task" [ 634.718646] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.734105] env[65107]: DEBUG oslo_vmware.api [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102270, 'name': ResetVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.748914] env[65107]: DEBUG oslo_concurrency.lockutils [req-783a1dfd-33c5-423f-bc7a-08fbe7f91389 req-b25330e9-b993-4890-98ef-26f6d04802da service nova] Releasing lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.826031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.828630] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.931560] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52753890-7bd2-599b-7f32-d5972d78c4e5, 'name': SearchDatastore_Task, 'duration_secs': 0.036308} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.931752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.932377] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 7a3009bf-54a2-4565-a1aa-1d19286a4810/7a3009bf-54a2-4565-a1aa-1d19286a4810.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 634.932377] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57fa9df7-0ddc-4c4e-82e9-87448d45e0ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.943963] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 634.943963] env[65107]: value = "task-5102271" [ 634.943963] env[65107]: _type = "Task" [ 634.943963] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.952238] env[65107]: DEBUG oslo_vmware.api [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102267, 'name': PowerOffVM_Task, 'duration_secs': 0.338343} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.952972] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 634.953277] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 634.953491] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab765dd6-8494-4691-a5fd-cfce044d9218 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.958546] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102271, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.976349] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-0691bdfe-ff23-471b-ad94-2a98adcecd4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.976870] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Instance network_info: |[{"id": "d9b6f78e-710c-4b4c-a9bd-04632f29764a", "address": "fa:16:3e:0c:fb:ab", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9b6f78e-71", "ovs_interfaceid": "d9b6f78e-710c-4b4c-a9bd-04632f29764a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 634.977481] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:fb:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9b6f78e-710c-4b4c-a9bd-04632f29764a', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.990324] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating folder: Project (4778fe9152224fd29d1f6220a19b5a36). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 634.991486] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcde444a-f1c8-4f6c-98d3-4675c9109f86 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.006110] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created folder: Project (4778fe9152224fd29d1f6220a19b5a36) in parent group-v992574. [ 635.006438] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating folder: Instances. Parent ref: group-v992619. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 635.006960] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-126e8580-cb46-4af5-83dd-e4df28992b4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.011228] env[65107]: DEBUG nova.scheduler.client.report [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 635.028314] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created folder: Instances in parent group-v992619. [ 635.028656] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 635.036358] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 635.037667] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c385e8b2-5745-4608-884f-315ce3cbb583 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.065444] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 635.065444] env[65107]: value = "task-5102275" [ 635.065444] env[65107]: _type = "Task" [ 635.065444] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.066520] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102268, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.067293] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 635.067554] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 635.067693] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Deleting the datastore file [datastore2] 0ef89aea-5373-4d07-9587-37890c08d35b {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 635.072189] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9d5369e-5939-4d9a-9295-862068a4cc0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.085606] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5219bc5c-97aa-d504-6b90-ba5e3d85c143, 'name': SearchDatastore_Task, 'duration_secs': 0.023529} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.090941] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.091362] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 635.092047] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.092047] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.092385] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 635.092681] env[65107]: DEBUG oslo_vmware.api [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for the task: (returnval){ [ 635.092681] env[65107]: value = "task-5102276" [ 635.092681] env[65107]: _type = "Task" [ 635.092681] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.093327] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102275, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.094060] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5eca4a1e-b2a5-457f-aa13-ff88fc19bc45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.096281] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Releasing lock "refresh_cache-abb9dc0e-ce18-49c8-a472-d39a3fcec887" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.096633] env[65107]: DEBUG nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Received event network-vif-plugged-b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 635.096901] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Acquiring lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.097239] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.097571] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.097857] env[65107]: DEBUG nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] No waiting events found dispatching network-vif-plugged-b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 635.098203] env[65107]: WARNING nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Received unexpected event network-vif-plugged-b41e2bb0-3a13-473d-ac19-0e593444f482 for instance with vm_state active and task_state None. [ 635.098441] env[65107]: DEBUG nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Received event network-changed-b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 635.098693] env[65107]: DEBUG nova.compute.manager [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Refreshing instance network info cache due to event network-changed-b41e2bb0-3a13-473d-ac19-0e593444f482. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 635.099062] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Acquiring lock "refresh_cache-eac4b8f9-9cd5-44a3-a12e-9ec22767b907" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.099190] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Acquired lock "refresh_cache-eac4b8f9-9cd5-44a3-a12e-9ec22767b907" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.099375] env[65107]: DEBUG nova.network.neutron [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Refreshing network info cache for port b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 635.114323] env[65107]: DEBUG oslo_vmware.api [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102276, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.126402] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 635.126402] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 635.127348] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1a36623-8964-40dd-b734-e66d730c2697 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.136964] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 635.136964] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b692d2-6416-2f11-1b51-b646ac3b8da5" [ 635.136964] env[65107]: _type = "Task" [ 635.136964] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.149277] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b692d2-6416-2f11-1b51-b646ac3b8da5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.160865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "refresh_cache-b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.161130] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired lock "refresh_cache-b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.161634] env[65107]: DEBUG nova.network.neutron [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 635.205045] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102266, 'name': CreateVM_Task, 'duration_secs': 0.726663} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.205226] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 635.207641] env[65107]: WARNING neutronclient.v2_0.client [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.208195] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.208364] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.209036] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 635.209147] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b05176d9-2626-4af4-934b-392b9300693b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.217805] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 635.217805] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52cbc438-76e8-118b-6371-203676066fee" [ 635.217805] env[65107]: _type = "Task" [ 635.217805] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.233733] env[65107]: DEBUG oslo_vmware.api [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102270, 'name': ResetVM_Task, 'duration_secs': 0.123552} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.238721] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Did hard reboot of VM {{(pid=65107) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 635.239250] env[65107]: DEBUG nova.compute.manager [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 635.239644] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52cbc438-76e8-118b-6371-203676066fee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.240776] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4aff6b0-71ba-4ee6-90ec-0f9540f96b8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.330736] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 635.460586] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102271, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.519048] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.467s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.519600] env[65107]: INFO nova.compute.manager [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Migrating [ 635.520310] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.520476] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "compute-rpcapi-router" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.522577] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.169s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.524790] env[65107]: INFO nova.compute.claims [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.543752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.544338] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.546542] env[65107]: DEBUG oslo_vmware.api [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Task: {'id': task-5102268, 'name': PowerOnVM_Task, 'duration_secs': 0.574625} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.547531] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 635.547926] env[65107]: DEBUG nova.compute.manager [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 635.549027] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b271a166-eafb-4688-afa9-414f2ad98ba2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.584915] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102275, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.608091] env[65107]: WARNING neutronclient.v2_0.client [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.608788] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.609235] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.616800] env[65107]: DEBUG oslo_vmware.api [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Task: {'id': task-5102276, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.472932} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.617711] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 635.617928] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 635.618080] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 635.618252] env[65107]: INFO nova.compute.manager [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Took 1.23 seconds to destroy the instance on the hypervisor. [ 635.618491] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 635.618998] env[65107]: DEBUG nova.compute.manager [-] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 635.618998] env[65107]: DEBUG nova.network.neutron [-] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 635.618998] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.619428] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.619675] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.648569] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b692d2-6416-2f11-1b51-b646ac3b8da5, 'name': SearchDatastore_Task, 'duration_secs': 0.014737} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.649356] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1d8b36-dd5d-4bf3-8e60-ad3cd7691e64 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.655620] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 635.655620] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52436ae3-9650-89a0-0178-cc6a3e8b86d9" [ 635.655620] env[65107]: _type = "Task" [ 635.655620] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.665443] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.665790] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.672455] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52436ae3-9650-89a0-0178-cc6a3e8b86d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.729081] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52cbc438-76e8-118b-6371-203676066fee, 'name': SearchDatastore_Task, 'duration_secs': 0.031852} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.729877] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.730342] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 635.730892] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.731240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.731604] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 635.732114] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b982f8aa-7e00-4da8-aff7-ef4d4f63ae96 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.744604] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 635.744604] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 635.745874] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff4430a2-5a86-4e91-83f3-9249d1c1b167 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.751522] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 635.751522] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b4d455-f54c-ea1a-2c4f-c44f8f7345a9" [ 635.751522] env[65107]: _type = "Task" [ 635.751522] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.758312] env[65107]: DEBUG oslo_concurrency.lockutils [None req-119801d8-34bb-471b-b441-d357f68ffa3f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 8.073s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.763276] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b4d455-f54c-ea1a-2c4f-c44f8f7345a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.867089] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.957209] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102271, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614529} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.957632] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 7a3009bf-54a2-4565-a1aa-1d19286a4810/7a3009bf-54a2-4565-a1aa-1d19286a4810.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.958436] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 635.958436] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2536fbc-d198-41ac-9edf-fd5678dd7c26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.967871] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 635.967871] env[65107]: value = "task-5102277" [ 635.967871] env[65107]: _type = "Task" [ 635.967871] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.979825] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102277, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.023577] env[65107]: INFO nova.compute.rpcapi [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 636.024450] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "compute-rpcapi-router" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.051022] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 636.078928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.086074] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102275, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.167384] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52436ae3-9650-89a0-0178-cc6a3e8b86d9, 'name': SearchDatastore_Task, 'duration_secs': 0.026371} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.167693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.168013] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] be518c1d-edd7-40ff-b7cc-3310885b07b4/be518c1d-edd7-40ff-b7cc-3310885b07b4.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.168339] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1702457f-69e4-43d2-9055-4d53c12d699f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.177314] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 636.177314] env[65107]: value = "task-5102278" [ 636.177314] env[65107]: _type = "Task" [ 636.177314] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.187224] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.236465] env[65107]: DEBUG nova.network.neutron [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 636.240129] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.262933] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b4d455-f54c-ea1a-2c4f-c44f8f7345a9, 'name': SearchDatastore_Task, 'duration_secs': 0.05875} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.263641] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaa3d0e0-06c4-4ff0-a0c9-9c86fb6ea2ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.270165] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 636.270165] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ff1dd9-554c-03aa-126e-ce9593b78d2e" [ 636.270165] env[65107]: _type = "Task" [ 636.270165] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.278838] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ff1dd9-554c-03aa-126e-ce9593b78d2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.479385] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102277, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080002} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.479553] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 636.481116] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693ef799-8306-4cb7-a268-9476fd06dc7e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.516249] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 7a3009bf-54a2-4565-a1aa-1d19286a4810/7a3009bf-54a2-4565-a1aa-1d19286a4810.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 636.516816] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3008f24-634b-4b6b-a374-5b8364afe15f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.541036] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 636.541036] env[65107]: value = "task-5102279" [ 636.541036] env[65107]: _type = "Task" [ 636.541036] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.551855] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.552147] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.553998] env[65107]: DEBUG nova.network.neutron [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 636.559188] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102279, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.581523] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.585332] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102275, 'name': CreateVM_Task, 'duration_secs': 1.073007} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.585622] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 636.586265] env[65107]: WARNING neutronclient.v2_0.client [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.586935] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.587174] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.587831] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 636.588631] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe20243d-5790-49da-b055-eb41c03b8b93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.597310] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 636.597310] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202d3f0-0c21-4fec-7b66-8b7720d49e28" [ 636.597310] env[65107]: _type = "Task" [ 636.597310] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.612208] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202d3f0-0c21-4fec-7b66-8b7720d49e28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.697946] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.792112] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ff1dd9-554c-03aa-126e-ce9593b78d2e, 'name': SearchDatastore_Task, 'duration_secs': 0.066066} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.798264] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.798264] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] f2814dd6-8be5-49b2-a030-f57e452c2e6d/f2814dd6-8be5-49b2-a030-f57e452c2e6d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.798581] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c77ec9ba-c8a5-4414-ab15-a0fe525bf78f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.812910] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 636.812910] env[65107]: value = "task-5102281" [ 636.812910] env[65107]: _type = "Task" [ 636.812910] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.831815] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102281, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.927659] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.928666] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.960855] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca8e1e9-2499-4615-b6fb-e0d798e6d809 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.970925] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970b3ee3-783c-43bc-a83a-eb87b0bcfcc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.008821] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db8d182-807f-4b61-b463-6c4d88b745c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.020814] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1db616d-0ad1-4171-a3ea-79a7ef109610 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.040718] env[65107]: DEBUG nova.compute.provider_tree [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.060331] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.061023] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.061534] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.070676] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102279, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.111594] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202d3f0-0c21-4fec-7b66-8b7720d49e28, 'name': SearchDatastore_Task, 'duration_secs': 0.036717} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.111986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.112300] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 637.112579] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.113105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.113105] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 637.113446] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae1b856d-d764-4436-b6f8-32ad9ac4a6f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.130154] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 637.130360] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 637.131870] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-826d30c0-5480-4537-ac88-f1bdc029878a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.141611] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 637.141611] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526ffb0c-f66b-da25-08eb-f5594c39beee" [ 637.141611] env[65107]: _type = "Task" [ 637.141611] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.156528] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526ffb0c-f66b-da25-08eb-f5594c39beee, 'name': SearchDatastore_Task, 'duration_secs': 0.012494} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.157617] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-371adf97-966c-4af3-a571-c712f4b8436c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.164635] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 637.164635] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f49b4f-10de-6113-578a-4523b871d34c" [ 637.164635] env[65107]: _type = "Task" [ 637.164635] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.175420] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f49b4f-10de-6113-578a-4523b871d34c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.193637] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.981371} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.193926] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] be518c1d-edd7-40ff-b7cc-3310885b07b4/be518c1d-edd7-40ff-b7cc-3310885b07b4.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 637.194207] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 637.194467] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e6af0f1-87ba-4825-9441-f6554a1f8595 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.203503] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.205144] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.213400] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 637.213400] env[65107]: value = "task-5102282" [ 637.213400] env[65107]: _type = "Task" [ 637.213400] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.224901] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102282, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.329807] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102281, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.550694] env[65107]: DEBUG nova.scheduler.client.report [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 637.565659] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102279, 'name': ReconfigVM_Task, 'duration_secs': 0.783985} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.566072] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 7a3009bf-54a2-4565-a1aa-1d19286a4810/7a3009bf-54a2-4565-a1aa-1d19286a4810.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 637.567152] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da36f170-3d77-45f2-9ed6-001343fe19ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.576546] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 637.576546] env[65107]: value = "task-5102283" [ 637.576546] env[65107]: _type = "Task" [ 637.576546] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.591039] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102283, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.683446] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f49b4f-10de-6113-578a-4523b871d34c, 'name': SearchDatastore_Task, 'duration_secs': 0.024047} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.686793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.687950] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0691bdfe-ff23-471b-ad94-2a98adcecd4f/0691bdfe-ff23-471b-ad94-2a98adcecd4f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 637.688365] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-758553ce-6892-4622-9ac6-07de6cff2f9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.698579] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 637.698579] env[65107]: value = "task-5102284" [ 637.698579] env[65107]: _type = "Task" [ 637.698579] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.709898] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102284, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.725554] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102282, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071725} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.725884] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 637.726720] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937a1814-2590-4cdc-82dd-342cf5553638 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.731595] env[65107]: WARNING neutronclient.v2_0.client [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.732319] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.732648] env[65107]: WARNING openstack [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.741517] env[65107]: DEBUG nova.network.neutron [-] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 637.769343] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] be518c1d-edd7-40ff-b7cc-3310885b07b4/be518c1d-edd7-40ff-b7cc-3310885b07b4.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 637.770444] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e13fcbc2-4242-43be-8ac2-8b81369bd45e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.786921] env[65107]: WARNING neutronclient.v2_0.client [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.787819] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.787971] env[65107]: WARNING openstack [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.813632] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 637.813632] env[65107]: value = "task-5102285" [ 637.813632] env[65107]: _type = "Task" [ 637.813632] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.825464] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102285, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.838782] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102281, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.858874} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.839102] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] f2814dd6-8be5-49b2-a030-f57e452c2e6d/f2814dd6-8be5-49b2-a030-f57e452c2e6d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 637.839275] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 637.839644] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e662ae62-67b8-44d0-8b2f-9cdcb604ba76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.851183] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 637.851183] env[65107]: value = "task-5102286" [ 637.851183] env[65107]: _type = "Task" [ 637.851183] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.862202] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.996241] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.999926] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.047754] env[65107]: DEBUG nova.network.neutron [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Updating instance_info_cache with network_info: [{"id": "f9582d02-fbd0-490d-8f90-10bc9cf32af8", "address": "fa:16:3e:e0:d5:e4", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9582d02-fb", "ovs_interfaceid": "f9582d02-fbd0-490d-8f90-10bc9cf32af8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.056889] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.057604] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 638.060346] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.194s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.062150] env[65107]: INFO nova.compute.claims [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.096701] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102283, 'name': Rename_Task, 'duration_secs': 0.325124} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.097565] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 638.098565] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3cdeebc-0f06-4d7b-bb35-001f212a20a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.111333] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 638.111333] env[65107]: value = "task-5102287" [ 638.111333] env[65107]: _type = "Task" [ 638.111333] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.124188] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.193888] env[65107]: DEBUG nova.network.neutron [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Updated VIF entry in instance network info cache for port b41e2bb0-3a13-473d-ac19-0e593444f482. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 638.195271] env[65107]: DEBUG nova.network.neutron [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Updating instance_info_cache with network_info: [{"id": "b41e2bb0-3a13-473d-ac19-0e593444f482", "address": "fa:16:3e:42:de:10", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb41e2bb0-3a", "ovs_interfaceid": "b41e2bb0-3a13-473d-ac19-0e593444f482", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.214614] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102284, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.229046] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.229923] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.230421] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.247743] env[65107]: INFO nova.compute.manager [-] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Took 2.63 seconds to deallocate network for instance. [ 638.331420] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.366884] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128756} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.367487] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.368615] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736bb4af-a6c6-46b7-9da3-171faf0a901d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.375806] env[65107]: DEBUG nova.network.neutron [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance_info_cache with network_info: [{"id": "086361c0-3cd7-4389-b5f0-c550a2adc834", "address": "fa:16:3e:bf:a0:a1", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086361c0-3c", "ovs_interfaceid": "086361c0-3cd7-4389-b5f0-c550a2adc834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.413516] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] f2814dd6-8be5-49b2-a030-f57e452c2e6d/f2814dd6-8be5-49b2-a030-f57e452c2e6d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.414424] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b3f230a-6869-44b9-9288-c95801150341 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.441855] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 638.441855] env[65107]: value = "task-5102288" [ 638.441855] env[65107]: _type = "Task" [ 638.441855] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.453080] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102288, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.462609] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Received event network-vif-plugged-0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.462609] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Acquiring lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.462609] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.462609] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.462609] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] No waiting events found dispatching network-vif-plugged-0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 638.462885] env[65107]: WARNING nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Received unexpected event network-vif-plugged-0b52bca6-ff8c-4322-9b18-1c1129186df6 for instance with vm_state building and task_state spawning. [ 638.462885] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Received event network-changed-0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.462885] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Refreshing instance network info cache due to event network-changed-0b52bca6-ff8c-4322-9b18-1c1129186df6. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 638.462885] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Acquiring lock "refresh_cache-f2814dd6-8be5-49b2-a030-f57e452c2e6d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.463040] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Acquired lock "refresh_cache-f2814dd6-8be5-49b2-a030-f57e452c2e6d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.463077] env[65107]: DEBUG nova.network.neutron [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Refreshing network info cache for port 0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 638.551445] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Received event network-vif-plugged-20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.551701] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquiring lock "7a3009bf-54a2-4565-a1aa-1d19286a4810-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.551915] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.552120] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.552308] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] No waiting events found dispatching network-vif-plugged-20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 638.552479] env[65107]: WARNING nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Received unexpected event network-vif-plugged-20915f30-cfbb-4a09-bf64-495545e0096f for instance with vm_state building and task_state spawning. [ 638.552652] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Received event network-changed-20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.552817] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Refreshing instance network info cache due to event network-changed-20915f30-cfbb-4a09-bf64-495545e0096f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 638.553024] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquiring lock "refresh_cache-7a3009bf-54a2-4565-a1aa-1d19286a4810" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.553196] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquired lock "refresh_cache-7a3009bf-54a2-4565-a1aa-1d19286a4810" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.553399] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Refreshing network info cache for port 20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 638.558794] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Releasing lock "refresh_cache-b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.559187] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Instance network_info: |[{"id": "f9582d02-fbd0-490d-8f90-10bc9cf32af8", "address": "fa:16:3e:e0:d5:e4", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9582d02-fb", "ovs_interfaceid": "f9582d02-fbd0-490d-8f90-10bc9cf32af8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 638.560058] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:d5:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9582d02-fbd0-490d-8f90-10bc9cf32af8', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 638.570786] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 638.574857] env[65107]: DEBUG nova.compute.utils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 638.581387] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 638.581626] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 638.581925] env[65107]: DEBUG nova.network.neutron [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 638.582314] env[65107]: WARNING neutronclient.v2_0.client [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.582655] env[65107]: WARNING neutronclient.v2_0.client [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.583378] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.583889] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.591228] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c470a12-3c90-4a48-9763-87b6b90884c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.618453] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 638.618453] env[65107]: value = "task-5102289" [ 638.618453] env[65107]: _type = "Task" [ 638.618453] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.626660] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102287, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.636722] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102289, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.681348] env[65107]: DEBUG nova.policy [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd7f9abea8d946ecb2accee2714703e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2d4844feda0432ebe1ef160b0070b83', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 638.698914] env[65107]: DEBUG oslo_concurrency.lockutils [req-a2d7837c-5047-4a69-93ed-9b5f03f08831 req-a266fbc0-e7f8-47d8-aedc-8c767a9ddc2a service nova] Releasing lock "refresh_cache-eac4b8f9-9cd5-44a3-a12e-9ec22767b907" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.712130] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102284, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.745164} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.712683] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0691bdfe-ff23-471b-ad94-2a98adcecd4f/0691bdfe-ff23-471b-ad94-2a98adcecd4f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 638.714871] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 638.715375] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca2d6bda-c136-4a5b-9c19-1aa8fad2ccaa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.726432] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 638.726432] env[65107]: value = "task-5102290" [ 638.726432] env[65107]: _type = "Task" [ 638.726432] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.742796] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102290, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.758121] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.830771] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102285, 'name': ReconfigVM_Task, 'duration_secs': 0.703961} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.831226] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Reconfigured VM instance instance-0000000d to attach disk [datastore1] be518c1d-edd7-40ff-b7cc-3310885b07b4/be518c1d-edd7-40ff-b7cc-3310885b07b4.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 638.832187] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a19bbc43-d009-446a-ac65-eba90c9a615d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.841743] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 638.841743] env[65107]: value = "task-5102291" [ 638.841743] env[65107]: _type = "Task" [ 638.841743] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.858138] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102291, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.880363] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.955863] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102288, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.966658] env[65107]: WARNING neutronclient.v2_0.client [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.967524] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.967862] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.061349] env[65107]: WARNING neutronclient.v2_0.client [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.062699] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.063426] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.082548] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 639.144342] env[65107]: DEBUG oslo_vmware.api [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102287, 'name': PowerOnVM_Task, 'duration_secs': 0.935343} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.145276] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102289, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.145276] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 639.145276] env[65107]: INFO nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Took 15.21 seconds to spawn the instance on the hypervisor. [ 639.145542] env[65107]: DEBUG nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 639.147320] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a48d03e-7862-4c47-8700-56a351ef521f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.238793] env[65107]: DEBUG nova.network.neutron [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Successfully created port: 321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 639.247409] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102290, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.390379} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.247707] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 639.248669] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb29e4b5-e82d-40e4-8a21-ec65e7a75481 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.284016] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 0691bdfe-ff23-471b-ad94-2a98adcecd4f/0691bdfe-ff23-471b-ad94-2a98adcecd4f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 639.284485] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2a692de-1b08-49cb-96dc-df26b27a5648 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.315594] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 639.315594] env[65107]: value = "task-5102293" [ 639.315594] env[65107]: _type = "Task" [ 639.315594] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.329150] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102293, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.358669] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102291, 'name': Rename_Task, 'duration_secs': 0.395785} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.359050] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 639.359333] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a1c9fe8-982e-472b-86e6-32b86b239e3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.368381] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 639.368381] env[65107]: value = "task-5102294" [ 639.368381] env[65107]: _type = "Task" [ 639.368381] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.381674] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.456836] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102288, 'name': ReconfigVM_Task, 'duration_secs': 0.851151} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.457157] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Reconfigured VM instance instance-0000000e to attach disk [datastore2] f2814dd6-8be5-49b2-a030-f57e452c2e6d/f2814dd6-8be5-49b2-a030-f57e452c2e6d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.457918] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cf5ed5a-39f0-4645-95e1-ee87bbf530f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.470321] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 639.470321] env[65107]: value = "task-5102295" [ 639.470321] env[65107]: _type = "Task" [ 639.470321] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.483047] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102295, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.513871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d74fb9-de89-4bcd-bbae-1b2185608436 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.522740] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c13f4a-9e66-496f-bb10-9f67f8282f79 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.559496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b81489d-9a3e-4aca-b8f0-6f7c9876a4eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.571525] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e0dd75-54b5-444a-87ea-6c76ba6489b4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.594019] env[65107]: DEBUG nova.compute.provider_tree [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.631798] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102289, 'name': CreateVM_Task, 'duration_secs': 0.722835} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.632048] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 639.632650] env[65107]: WARNING neutronclient.v2_0.client [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.633169] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.633327] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.633662] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 639.633958] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-514dcefd-68c4-47e3-a22f-c173e49e0c07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.640208] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 639.640208] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5201f2a9-1d26-bdd1-754e-f5428f55d57a" [ 639.640208] env[65107]: _type = "Task" [ 639.640208] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.650849] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5201f2a9-1d26-bdd1-754e-f5428f55d57a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.685374] env[65107]: INFO nova.compute.manager [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Took 20.19 seconds to build instance. [ 639.829750] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102293, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.881963] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102294, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.982260] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102295, 'name': Rename_Task, 'duration_secs': 0.360095} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.982454] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 639.982691] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-832b0d26-556d-4492-9991-fa45521a3423 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.990731] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 639.990731] env[65107]: value = "task-5102296" [ 639.990731] env[65107]: _type = "Task" [ 639.990731] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.001527] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102296, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.100183] env[65107]: DEBUG nova.scheduler.client.report [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 640.105259] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 640.133274] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 640.133582] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 640.133749] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 640.134037] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 640.134198] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 640.134348] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 640.134557] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.134710] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 640.134869] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 640.135128] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 640.135288] env[65107]: DEBUG nova.virt.hardware [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 640.136536] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477ce013-3293-4b9c-9eae-b24f2f3ea699 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.148672] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1693b0e4-9c03-4bb7-aa50-0f4a870dd294 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.157051] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5201f2a9-1d26-bdd1-754e-f5428f55d57a, 'name': SearchDatastore_Task, 'duration_secs': 0.054335} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.157804] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.158089] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.158398] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.158589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.158779] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.159088] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae980631-7e76-43a3-a7e4-301b73c76b4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.184925] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.185184] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 640.185938] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b5b9ce6-7c96-44ed-ad32-ee491651edbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.188840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9d118051-74b2-4029-8ac3-3410116a4ac9 tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.699s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.193799] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 640.193799] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d9787b-4daa-4ba8-cde2-ebf89f8f88f3" [ 640.193799] env[65107]: _type = "Task" [ 640.193799] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.204894] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d9787b-4daa-4ba8-cde2-ebf89f8f88f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.327120] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102293, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.383109] env[65107]: DEBUG oslo_vmware.api [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102294, 'name': PowerOnVM_Task, 'duration_secs': 0.8644} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.383473] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 640.383740] env[65107]: INFO nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Took 14.00 seconds to spawn the instance on the hypervisor. [ 640.383997] env[65107]: DEBUG nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 640.385188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad2a731-3994-4586-b99b-35e47df1d6ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.402264] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e7b52f-64c0-41d9-8583-475aee502530 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.425544] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 640.501187] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102296, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.610389] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.611210] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 640.615599] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.537s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.615910] env[65107]: DEBUG nova.objects.instance [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 640.706390] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d9787b-4daa-4ba8-cde2-ebf89f8f88f3, 'name': SearchDatastore_Task, 'duration_secs': 0.014006} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.707246] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21a7a74b-f91f-43a2-aabe-05532efea012 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.714339] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 640.714339] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522b632b-d3de-d288-10cd-c2681e8687ef" [ 640.714339] env[65107]: _type = "Task" [ 640.714339] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.725178] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522b632b-d3de-d288-10cd-c2681e8687ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.834437] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102293, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.881449] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.881977] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.896576] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.897057] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.916702] env[65107]: INFO nova.compute.manager [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Took 19.94 seconds to build instance. [ 640.934066] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 640.934129] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a87767c-f442-4c81-b6c5-8b106a22480e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.943183] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 640.943183] env[65107]: value = "task-5102297" [ 640.943183] env[65107]: _type = "Task" [ 640.943183] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.952095] env[65107]: DEBUG nova.network.neutron [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Successfully updated port: 321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 640.960950] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102297, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.010281] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102296, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.122187] env[65107]: DEBUG nova.compute.utils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 641.127486] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 641.127778] env[65107]: DEBUG nova.network.neutron [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 641.128236] env[65107]: WARNING neutronclient.v2_0.client [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.128632] env[65107]: WARNING neutronclient.v2_0.client [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.129354] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.129803] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.224517] env[65107]: WARNING neutronclient.v2_0.client [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.225682] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.226112] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.243260] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522b632b-d3de-d288-10cd-c2681e8687ef, 'name': SearchDatastore_Task, 'duration_secs': 0.015938} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.243537] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.244069] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8/b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 641.244526] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f772f0f6-e47f-47b4-af52-f354754fa15c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.248451] env[65107]: WARNING neutronclient.v2_0.client [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.249745] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.250493] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.266282] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 641.266282] env[65107]: value = "task-5102298" [ 641.266282] env[65107]: _type = "Task" [ 641.266282] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.276290] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102298, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.328776] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102293, 'name': ReconfigVM_Task, 'duration_secs': 1.619678} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.329254] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 0691bdfe-ff23-471b-ad94-2a98adcecd4f/0691bdfe-ff23-471b-ad94-2a98adcecd4f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 641.330043] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-561e2d77-fa88-4075-adf2-dbfc434c3a66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.338415] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 641.338415] env[65107]: value = "task-5102299" [ 641.338415] env[65107]: _type = "Task" [ 641.338415] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.351156] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102299, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.396272] env[65107]: DEBUG nova.policy [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2baf466e51274c5bbd1393e549939e7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bcb6424259344f3fa87ba9a4c10637d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 641.419398] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b289553a-89d6-4cb3-a9eb-b9f1edd283c9 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.459s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.454615] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102297, 'name': PowerOffVM_Task, 'duration_secs': 0.259324} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.454615] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 641.454782] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 641.458888] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "refresh_cache-f58c3088-c821-4eec-be0a-81221debc98e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.459087] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "refresh_cache-f58c3088-c821-4eec-be0a-81221debc98e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.459225] env[65107]: DEBUG nova.network.neutron [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 641.505952] env[65107]: DEBUG oslo_vmware.api [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102296, 'name': PowerOnVM_Task, 'duration_secs': 1.279079} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.506353] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 641.506568] env[65107]: INFO nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Took 13.24 seconds to spawn the instance on the hypervisor. [ 641.506750] env[65107]: DEBUG nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 641.507683] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a700990b-f66c-4a10-8dea-1fc945afde84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.646087] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178456f3-18bb-4960-b16d-1cbdb43a4af5 tempest-ServersAdmin275Test-1613043799 tempest-ServersAdmin275Test-1613043799-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.027s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.647196] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 641.652146] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.071s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.657084] env[65107]: INFO nova.compute.claims [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.778883] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102298, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.856871] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102299, 'name': Rename_Task, 'duration_secs': 0.191052} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.857369] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 641.859290] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9deb30b-40e0-45ea-ba66-c530e60eae0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.868030] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 641.868030] env[65107]: value = "task-5102300" [ 641.868030] env[65107]: _type = "Task" [ 641.868030] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.882726] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.964773] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 641.964773] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 641.965306] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 641.968133] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 641.968133] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 641.968133] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 641.968133] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.968133] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 641.968438] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 641.968438] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 641.968438] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 641.975187] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.975358] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.983178] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13d139c3-40be-4ffe-b6cd-301532054e62 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.001664] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 642.001664] env[65107]: value = "task-5102301" [ 642.001664] env[65107]: _type = "Task" [ 642.001664] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.012551] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102301, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.028696] env[65107]: DEBUG nova.network.neutron [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Updated VIF entry in instance network info cache for port 0b52bca6-ff8c-4322-9b18-1c1129186df6. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 642.029123] env[65107]: DEBUG nova.network.neutron [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Updating instance_info_cache with network_info: [{"id": "0b52bca6-ff8c-4322-9b18-1c1129186df6", "address": "fa:16:3e:64:fc:90", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b52bca6-ff", "ovs_interfaceid": "0b52bca6-ff8c-4322-9b18-1c1129186df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.036439] env[65107]: INFO nova.compute.manager [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Took 19.99 seconds to build instance. [ 642.066829] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Updated VIF entry in instance network info cache for port 20915f30-cfbb-4a09-bf64-495545e0096f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 642.067416] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Updating instance_info_cache with network_info: [{"id": "20915f30-cfbb-4a09-bf64-495545e0096f", "address": "fa:16:3e:e6:94:75", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.54", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20915f30-cf", "ovs_interfaceid": "20915f30-cfbb-4a09-bf64-495545e0096f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.120573] env[65107]: DEBUG nova.network.neutron [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 642.172596] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.172996] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.279633] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102298, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.671269} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.281131] env[65107]: WARNING neutronclient.v2_0.client [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.281750] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.282103] env[65107]: WARNING openstack [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.298034] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8/b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 642.298412] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.299320] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd7c6075-9a64-435e-8425-2b14cd80bfda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.308083] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 642.308083] env[65107]: value = "task-5102302" [ 642.308083] env[65107]: _type = "Task" [ 642.308083] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.327287] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102302, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.382136] env[65107]: DEBUG nova.network.neutron [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Successfully created port: fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 642.391705] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102300, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.439268] env[65107]: DEBUG nova.network.neutron [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Updating instance_info_cache with network_info: [{"id": "321da8b7-0ae1-438d-816f-6c711ef9b003", "address": "fa:16:3e:b6:63:49", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap321da8b7-0a", "ovs_interfaceid": "321da8b7-0ae1-438d-816f-6c711ef9b003", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.519117] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102301, 'name': ReconfigVM_Task, 'duration_secs': 0.231232} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.519472] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 642.534177] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Releasing lock "refresh_cache-f2814dd6-8be5-49b2-a030-f57e452c2e6d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.534454] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Received event network-vif-plugged-f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.534649] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Acquiring lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.534848] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.535014] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.535262] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] No waiting events found dispatching network-vif-plugged-f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 642.535446] env[65107]: WARNING nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Received unexpected event network-vif-plugged-f9582d02-fbd0-490d-8f90-10bc9cf32af8 for instance with vm_state building and task_state spawning. [ 642.535608] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Received event network-changed-f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.535763] env[65107]: DEBUG nova.compute.manager [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Refreshing instance network info cache due to event network-changed-f9582d02-fbd0-490d-8f90-10bc9cf32af8. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 642.535943] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Acquiring lock "refresh_cache-b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.536091] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Acquired lock "refresh_cache-b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.536251] env[65107]: DEBUG nova.network.neutron [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Refreshing network info cache for port f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 642.540121] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12ff576d-219d-4d86-8d14-68bddbc0047f tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.515s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.574342] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Releasing lock "refresh_cache-7a3009bf-54a2-4565-a1aa-1d19286a4810" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.574342] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Received event network-vif-deleted-7518822c-68c7-4bd3-8296-e61bd76671e2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.574342] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Received event network-vif-plugged-3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.574342] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquiring lock "be518c1d-edd7-40ff-b7cc-3310885b07b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.574342] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.574631] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.574714] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] No waiting events found dispatching network-vif-plugged-3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 642.574965] env[65107]: WARNING nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Received unexpected event network-vif-plugged-3bb17344-0031-4226-9117-1366ecf93330 for instance with vm_state building and task_state spawning. [ 642.575198] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Received event network-changed-3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.575530] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Refreshing instance network info cache due to event network-changed-3bb17344-0031-4226-9117-1366ecf93330. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 642.575820] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquiring lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.576042] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquired lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.576329] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Refreshing network info cache for port 3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 642.670472] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 642.690706] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.691055] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.691279] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.691540] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.691616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.694413] env[65107]: INFO nova.compute.manager [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Terminating instance [ 642.710391] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 642.710676] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 642.711276] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 642.711276] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 642.711276] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 642.711448] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 642.711680] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.711813] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 642.712024] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 642.712303] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 642.712303] env[65107]: DEBUG nova.virt.hardware [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 642.713704] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f0d25d-4130-47c6-a562-c06f6d0bb3b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.727332] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bebaf8e-35cf-4269-9a16-2022cfc3569c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.831074] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102302, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.135874} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.831074] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 642.831645] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8c272f-9e16-4154-9309-454f5c8740a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.863215] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8/b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 642.868954] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-796eb18e-3535-410b-b5a0-fbf7921147eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.903201] env[65107]: DEBUG oslo_vmware.api [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102300, 'name': PowerOnVM_Task, 'duration_secs': 0.755255} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.903579] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 642.903579] env[65107]: value = "task-5102303" [ 642.903579] env[65107]: _type = "Task" [ 642.903579] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.903906] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 642.904174] env[65107]: INFO nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Took 12.19 seconds to spawn the instance on the hypervisor. [ 642.904420] env[65107]: DEBUG nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 642.909018] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4da2d5b-e84f-4de4-9f61-32281d0386fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.925992] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102303, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.943136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "refresh_cache-f58c3088-c821-4eec-be0a-81221debc98e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.943561] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Instance network_info: |[{"id": "321da8b7-0ae1-438d-816f-6c711ef9b003", "address": "fa:16:3e:b6:63:49", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap321da8b7-0a", "ovs_interfaceid": "321da8b7-0ae1-438d-816f-6c711ef9b003", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 642.944190] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:63:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '321da8b7-0ae1-438d-816f-6c711ef9b003', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 642.952956] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 642.955882] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 642.956618] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23a6a6f6-1cf9-446b-9bc2-19684bf25142 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.982651] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 642.982651] env[65107]: value = "task-5102304" [ 642.982651] env[65107]: _type = "Task" [ 642.982651] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.996150] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102304, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.027138] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 643.027502] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 643.027638] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 643.027861] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 643.028084] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 643.028307] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 643.028553] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.028709] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 643.028884] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 643.029051] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 643.029226] env[65107]: DEBUG nova.virt.hardware [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 643.035223] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Reconfiguring VM instance instance-00000006 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 643.041512] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c71822b-ae47-4787-9cca-0d80515332ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.056662] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.056871] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.058572] env[65107]: WARNING neutronclient.v2_0.client [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.059286] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.059713] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.077878] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 643.077878] env[65107]: value = "task-5102305" [ 643.077878] env[65107]: _type = "Task" [ 643.077878] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.081674] env[65107]: WARNING neutronclient.v2_0.client [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.082383] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.082769] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.100900] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.127342] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e3832d-fcb6-4a79-a2aa-c8962a563b5c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.138532] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7b06be-8f99-4371-b19a-cf3fa6937326 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.172503] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79013903-181a-43f1-989f-559cc030480c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.187064] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6f10e5-8f76-4be3-b1a5-fe0f384d9f79 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.206473] env[65107]: DEBUG nova.compute.manager [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 643.206809] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.207753] env[65107]: DEBUG nova.compute.provider_tree [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.212899] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036785e9-fb3d-4100-9f16-42e1c9611c59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.225900] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 643.226782] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99d04d1d-ee1d-4de2-8043-5577fbc7448e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.237966] env[65107]: DEBUG oslo_vmware.api [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 643.237966] env[65107]: value = "task-5102306" [ 643.237966] env[65107]: _type = "Task" [ 643.237966] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.248414] env[65107]: DEBUG oslo_vmware.api [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102306, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.421374] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102303, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.440278] env[65107]: INFO nova.compute.manager [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Took 20.32 seconds to build instance. [ 643.494680] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102304, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.570229] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 643.595033] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102305, 'name': ReconfigVM_Task, 'duration_secs': 0.284166} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.595870] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Reconfigured VM instance instance-00000006 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 643.596952] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46090c9e-7910-4c33-9621-b75968c0b0a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.623650] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311/fdb5603d-24b0-4fdd-b8cc-25c0ebee5311.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.624424] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fa788fa-099f-4ad5-b879-f599b3572de8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.647876] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 643.647876] env[65107]: value = "task-5102307" [ 643.647876] env[65107]: _type = "Task" [ 643.647876] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.659763] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102307, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.720977] env[65107]: DEBUG nova.scheduler.client.report [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 643.750682] env[65107]: DEBUG oslo_vmware.api [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102306, 'name': PowerOffVM_Task, 'duration_secs': 0.47001} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.751238] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 643.751441] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 643.751711] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4aab9563-f0b9-4563-907a-e618ed2b6dff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.866931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "7b24c530-56bf-4666-96f9-e2dc66728f1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.867215] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.922450] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102303, 'name': ReconfigVM_Task, 'duration_secs': 0.834436} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.922978] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Reconfigured VM instance instance-00000010 to attach disk [datastore2] b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8/b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.923843] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6134456c-2c8f-4d14-9da9-5cb247a4d251 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.933632] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 643.933632] env[65107]: value = "task-5102309" [ 643.933632] env[65107]: _type = "Task" [ 643.933632] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.948795] env[65107]: DEBUG oslo_concurrency.lockutils [None req-347b3519-af5b-4732-8546-3508f3c4a7a2 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.837s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.949398] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102309, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.984373] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 643.984746] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 643.984804] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleting the datastore file [datastore1] eac4b8f9-9cd5-44a3-a12e-9ec22767b907 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.988512] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b9522f3-2151-4c57-a88c-d61b5d4eb331 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.998195] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102304, 'name': CreateVM_Task, 'duration_secs': 0.633298} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.000248] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 644.000725] env[65107]: DEBUG oslo_vmware.api [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 644.000725] env[65107]: value = "task-5102310" [ 644.000725] env[65107]: _type = "Task" [ 644.000725] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.001361] env[65107]: WARNING neutronclient.v2_0.client [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.001809] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.002019] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.002411] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 644.002843] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd4abe23-7236-4c49-995a-4f3e1d454df6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.012853] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 644.012853] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fed71-f67e-ed13-070b-6b64f1808cb0" [ 644.012853] env[65107]: _type = "Task" [ 644.012853] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.017706] env[65107]: DEBUG oslo_vmware.api [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102310, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.028473] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fed71-f67e-ed13-070b-6b64f1808cb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.101819] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.155698] env[65107]: DEBUG nova.network.neutron [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Successfully updated port: fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 644.162340] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102307, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.226508] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.227492] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 644.232738] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.475s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.233673] env[65107]: DEBUG nova.objects.instance [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lazy-loading 'resources' on Instance uuid 0ef89aea-5373-4d07-9587-37890c08d35b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 644.342512] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.342655] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.370484] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 644.398740] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.398740] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.447347] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102309, 'name': Rename_Task, 'duration_secs': 0.201421} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.448106] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.448386] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5aeffe44-8064-4b21-ad34-6c79416beea8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.456548] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 644.456548] env[65107]: value = "task-5102311" [ 644.456548] env[65107]: _type = "Task" [ 644.456548] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.468106] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102311, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.514514] env[65107]: DEBUG oslo_vmware.api [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102310, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377052} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.514867] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.515023] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 644.515218] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.515380] env[65107]: INFO nova.compute.manager [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Took 1.31 seconds to destroy the instance on the hypervisor. [ 644.515631] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 644.515951] env[65107]: DEBUG nova.compute.manager [-] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 644.516111] env[65107]: DEBUG nova.network.neutron [-] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 644.516646] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.517455] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.517769] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.537512] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fed71-f67e-ed13-070b-6b64f1808cb0, 'name': SearchDatastore_Task, 'duration_secs': 0.033336} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.537888] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 644.538262] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 644.538506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.538954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.538954] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 644.539228] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c03e0bdb-b603-47a7-9bc2-d4f3145b9aad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.550765] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 644.551050] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 644.552077] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36fca37e-91dc-4523-bbc0-1e27341c171f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.562698] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 644.562698] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52aa7f51-c49f-e2ad-5efa-a870b773c500" [ 644.562698] env[65107]: _type = "Task" [ 644.562698] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.572854] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52aa7f51-c49f-e2ad-5efa-a870b773c500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.660372] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102307, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.663997] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.664248] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.664385] env[65107]: DEBUG nova.network.neutron [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 644.737177] env[65107]: DEBUG nova.compute.utils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 644.742021] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 644.742021] env[65107]: DEBUG nova.network.neutron [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 644.743587] env[65107]: WARNING neutronclient.v2_0.client [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.743587] env[65107]: WARNING neutronclient.v2_0.client [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.743587] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.743969] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.911565] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.968966] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102311, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.002924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.003343] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.071408] env[65107]: WARNING neutronclient.v2_0.client [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.072064] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.072490] env[65107]: WARNING openstack [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.085906] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52aa7f51-c49f-e2ad-5efa-a870b773c500, 'name': SearchDatastore_Task, 'duration_secs': 0.014261} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.090760] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a8b349c-7ae8-4676-9e47-baf038574678 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.097158] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 645.097158] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5264f6f2-502f-e2d5-7b10-e68ad8b00da7" [ 645.097158] env[65107]: _type = "Task" [ 645.097158] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.110848] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5264f6f2-502f-e2d5-7b10-e68ad8b00da7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.164676] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102307, 'name': ReconfigVM_Task, 'duration_secs': 1.468219} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.165380] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Reconfigured VM instance instance-00000006 to attach disk [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311/fdb5603d-24b0-4fdd-b8cc-25c0ebee5311.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 645.166635] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 645.178028] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.178028] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.194448] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.243242] env[65107]: DEBUG nova.policy [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad5877ef7ac74b1ebf503221370828f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '520207ef29e64aa799841fae36e469f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 645.251618] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 645.254591] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3d512d-df04-46f1-9c39-e2336938b5fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.259362] env[65107]: WARNING neutronclient.v2_0.client [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.260112] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.260461] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.275747] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2a4f25-1686-46e8-8398-f987cdc024ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.321639] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dbc638-1ba6-4ddf-ab6a-8c9e10236956 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.333056] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8398b9f2-352a-4570-b21e-e80c4c09e1b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.339305] env[65107]: DEBUG nova.network.neutron [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 645.356742] env[65107]: DEBUG nova.compute.provider_tree [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.469101] env[65107]: DEBUG oslo_vmware.api [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102311, 'name': PowerOnVM_Task, 'duration_secs': 0.802277} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.469396] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.469630] env[65107]: INFO nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Took 12.28 seconds to spawn the instance on the hypervisor. [ 645.469876] env[65107]: DEBUG nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 645.470709] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd81693b-2484-4a1f-aa34-4de5b2ab435f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.505520] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 645.608963] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5264f6f2-502f-e2d5-7b10-e68ad8b00da7, 'name': SearchDatastore_Task, 'duration_secs': 0.018063} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.609409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.609580] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] f58c3088-c821-4eec-be0a-81221debc98e/f58c3088-c821-4eec-be0a-81221debc98e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 645.611145] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17e769a4-9ff2-4345-a088-f7aef54361ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.620076] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 645.620076] env[65107]: value = "task-5102312" [ 645.620076] env[65107]: _type = "Task" [ 645.620076] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.630394] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102312, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.691643] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a165f49-6579-4b49-b8e5-8e3b71564db8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.717178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05431d83-b337-47db-91bf-3b6be2a65f8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.737089] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 645.749499] env[65107]: DEBUG nova.network.neutron [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Updated VIF entry in instance network info cache for port f9582d02-fbd0-490d-8f90-10bc9cf32af8. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 645.749853] env[65107]: DEBUG nova.network.neutron [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Updating instance_info_cache with network_info: [{"id": "f9582d02-fbd0-490d-8f90-10bc9cf32af8", "address": "fa:16:3e:e0:d5:e4", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9582d02-fb", "ovs_interfaceid": "f9582d02-fbd0-490d-8f90-10bc9cf32af8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 645.760076] env[65107]: DEBUG nova.network.neutron [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Successfully created port: 2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 645.765997] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.766427] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.828265] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Updated VIF entry in instance network info cache for port 3bb17344-0031-4226-9117-1366ecf93330. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 645.828468] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Updating instance_info_cache with network_info: [{"id": "3bb17344-0031-4226-9117-1366ecf93330", "address": "fa:16:3e:06:8b:e4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bb17344-00", "ovs_interfaceid": "3bb17344-0031-4226-9117-1366ecf93330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 645.860391] env[65107]: DEBUG nova.scheduler.client.report [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 645.989661] env[65107]: WARNING neutronclient.v2_0.client [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.991747] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.992830] env[65107]: WARNING openstack [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.008370] env[65107]: INFO nova.compute.manager [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Took 17.39 seconds to build instance. [ 646.047927] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.135763] env[65107]: DEBUG nova.network.neutron [-] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.137149] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102312, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.195511] env[65107]: DEBUG nova.network.neutron [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.243592] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.244033] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.252602] env[65107]: DEBUG oslo_concurrency.lockutils [req-e6197e06-6310-4b9a-a709-fff58d036252 req-6515bd58-d268-4c9f-81cf-a4a2125a4c6b service nova] Releasing lock "refresh_cache-b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.275867] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 646.316444] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 646.316729] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 646.316871] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 646.317082] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 646.317236] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 646.317386] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 646.317624] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.317806] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 646.318079] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 646.318292] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 646.318593] env[65107]: DEBUG nova.virt.hardware [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 646.319529] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2176c3d4-ea60-4d7d-bbe9-6234641212d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.331221] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Releasing lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.333029] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Received event network-vif-plugged-d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 646.333029] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquiring lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.333029] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.333029] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.333029] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] No waiting events found dispatching network-vif-plugged-d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 646.333980] env[65107]: WARNING nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Received unexpected event network-vif-plugged-d9b6f78e-710c-4b4c-a9bd-04632f29764a for instance with vm_state building and task_state spawning. [ 646.333980] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Received event network-changed-d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 646.333980] env[65107]: DEBUG nova.compute.manager [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Refreshing instance network info cache due to event network-changed-d9b6f78e-710c-4b4c-a9bd-04632f29764a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 646.333980] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquiring lock "refresh_cache-0691bdfe-ff23-471b-ad94-2a98adcecd4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.333980] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Acquired lock "refresh_cache-0691bdfe-ff23-471b-ad94-2a98adcecd4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.334527] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Refreshing network info cache for port d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 646.336377] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a82c67f-367d-42a2-8f73-e952196d9fa6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.358144] env[65107]: DEBUG nova.network.neutron [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Port 086361c0-3cd7-4389-b5f0-c550a2adc834 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 646.371764] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.139s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.376898] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.274s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.378372] env[65107]: INFO nova.compute.claims [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.428510] env[65107]: INFO nova.scheduler.client.report [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Deleted allocations for instance 0ef89aea-5373-4d07-9587-37890c08d35b [ 646.511543] env[65107]: DEBUG oslo_concurrency.lockutils [None req-351a9ca7-cd31-4f24-88d8-6cf08e54ac41 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.936s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.631823] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102312, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.75469} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.632085] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] f58c3088-c821-4eec-be0a-81221debc98e/f58c3088-c821-4eec-be0a-81221debc98e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 646.632295] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 646.632560] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7377d71-9cb0-4156-be4e-136a4f5a0561 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.640538] env[65107]: INFO nova.compute.manager [-] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Took 2.12 seconds to deallocate network for instance. [ 646.641459] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 646.641459] env[65107]: value = "task-5102313" [ 646.641459] env[65107]: _type = "Task" [ 646.641459] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.656357] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102313, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.701708] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.702199] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Instance network_info: |[{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 646.702712] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:7d:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa708f72-f0b6-4ac5-a002-0d6589e8ab0e', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 646.712590] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Creating folder: Project (bcb6424259344f3fa87ba9a4c10637d2). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 646.712997] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f693c9ac-57a2-42a8-9691-1aec246037e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.724537] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Created folder: Project (bcb6424259344f3fa87ba9a4c10637d2) in parent group-v992574. [ 646.724743] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Creating folder: Instances. Parent ref: group-v992624. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 646.724998] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bc3b762-ed14-4d72-86e3-edc52d8d183d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.736366] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Created folder: Instances in parent group-v992624. [ 646.736675] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 646.736896] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 646.737204] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-403c6c6b-bbd5-4f89-8511-0a898b036693 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.761031] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 646.761031] env[65107]: value = "task-5102316" [ 646.761031] env[65107]: _type = "Task" [ 646.761031] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.769713] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102316, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.785038] env[65107]: DEBUG nova.compute.manager [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Received event network-changed-e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 646.785261] env[65107]: DEBUG nova.compute.manager [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Refreshing instance network info cache due to event network-changed-e41be772-a45d-410b-acaf-09d30f9afa9d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 646.785501] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Acquiring lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.785683] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Acquired lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.787445] env[65107]: DEBUG nova.network.neutron [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Refreshing network info cache for port e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 646.842033] env[65107]: WARNING neutronclient.v2_0.client [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.843288] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.843817] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.889817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.890042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.890712] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.890712] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.890896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.894026] env[65107]: INFO nova.compute.manager [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Terminating instance [ 646.939315] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9991a720-83bc-4436-aa47-e042da250ab7 tempest-ServerDiagnosticsNegativeTest-1730136242 tempest-ServerDiagnosticsNegativeTest-1730136242-project-member] Lock "0ef89aea-5373-4d07-9587-37890c08d35b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.064s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 647.157014] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.157014] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102313, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.223526} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.157014] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 647.158023] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc04db30-d4bb-47d3-80c6-f22eee817820 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.164620] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.165047] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.193482] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] f58c3088-c821-4eec-be0a-81221debc98e/f58c3088-c821-4eec-be0a-81221debc98e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 647.193482] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-668975d7-045b-49ea-84e2-9bb578ddc762 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.214141] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 647.214141] env[65107]: value = "task-5102317" [ 647.214141] env[65107]: _type = "Task" [ 647.214141] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.223489] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.274214] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102316, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.289148] env[65107]: WARNING neutronclient.v2_0.client [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.289998] env[65107]: WARNING openstack [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.290513] env[65107]: WARNING openstack [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.392613] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.392812] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.392968] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 647.399833] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "refresh_cache-ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.399833] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquired lock "refresh_cache-ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.400919] env[65107]: DEBUG nova.network.neutron [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 647.529825] env[65107]: DEBUG nova.network.neutron [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Successfully updated port: 2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 647.564379] env[65107]: WARNING neutronclient.v2_0.client [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.567872] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.568443] env[65107]: WARNING openstack [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.656270] env[65107]: WARNING openstack [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.656663] env[65107]: WARNING openstack [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.707767] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Updated VIF entry in instance network info cache for port d9b6f78e-710c-4b4c-a9bd-04632f29764a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 647.708190] env[65107]: DEBUG nova.network.neutron [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Updating instance_info_cache with network_info: [{"id": "d9b6f78e-710c-4b4c-a9bd-04632f29764a", "address": "fa:16:3e:0c:fb:ab", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9b6f78e-71", "ovs_interfaceid": "d9b6f78e-710c-4b4c-a9bd-04632f29764a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 647.735309] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.781948] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102316, 'name': CreateVM_Task, 'duration_secs': 0.792859} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.783393] env[65107]: WARNING neutronclient.v2_0.client [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.784291] env[65107]: WARNING openstack [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.784534] env[65107]: WARNING openstack [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.792400] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 647.795973] env[65107]: WARNING neutronclient.v2_0.client [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.796550] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.796550] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.796768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 647.797330] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e1adb1d-07a8-425f-883c-ef5b2d3f307a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.803382] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 647.803382] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b27b56-b5a3-d91b-5eb6-6f2dee32205c" [ 647.803382] env[65107]: _type = "Task" [ 647.803382] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.814253] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b27b56-b5a3-d91b-5eb6-6f2dee32205c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.816265] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4ea030-25c5-4ed2-9aa8-f9eda16626da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.825523] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd5f328-c894-4158-819b-adbcca5c654b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.871056] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5694c3-de40-434a-9573-c872587f2a4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.881396] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ddb81a-daf7-456c-a49e-e3ecd94cfbf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.897944] env[65107]: DEBUG nova.compute.provider_tree [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.904259] env[65107]: WARNING neutronclient.v2_0.client [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.905021] env[65107]: WARNING openstack [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.905431] env[65107]: WARNING openstack [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.919320] env[65107]: DEBUG nova.network.neutron [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updated VIF entry in instance network info cache for port e41be772-a45d-410b-acaf-09d30f9afa9d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 647.919695] env[65107]: DEBUG nova.network.neutron [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updating instance_info_cache with network_info: [{"id": "e41be772-a45d-410b-acaf-09d30f9afa9d", "address": "fa:16:3e:f7:84:8e", "network": {"id": "9a766a32-cd65-4a5e-b338-308b5e640847", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-2025951181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d98fae2faa0a40c798e251737ccb31c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41be772-a4", "ovs_interfaceid": "e41be772-a45d-410b-acaf-09d30f9afa9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 647.932952] env[65107]: DEBUG nova.network.neutron [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 648.006758] env[65107]: DEBUG nova.network.neutron [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 648.032058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "refresh_cache-e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.032817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquired lock "refresh_cache-e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.032817] env[65107]: DEBUG nova.network.neutron [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 648.212705] env[65107]: DEBUG oslo_concurrency.lockutils [req-82481eb8-81b4-457b-aa8a-621602ae2d23 req-3cce3056-9a63-4866-b92c-b4e4d8de6b3a service nova] Releasing lock "refresh_cache-0691bdfe-ff23-471b-ad94-2a98adcecd4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.228728] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102317, 'name': ReconfigVM_Task, 'duration_secs': 0.693208} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.228728] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Reconfigured VM instance instance-00000011 to attach disk [datastore2] f58c3088-c821-4eec-be0a-81221debc98e/f58c3088-c821-4eec-be0a-81221debc98e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 648.229394] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77927fed-56d1-414b-bb45-7f52d1655e1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.238665] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 648.238665] env[65107]: value = "task-5102318" [ 648.238665] env[65107]: _type = "Task" [ 648.238665] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.254454] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102318, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.318879] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b27b56-b5a3-d91b-5eb6-6f2dee32205c, 'name': SearchDatastore_Task, 'duration_secs': 0.012831} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.320437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.320690] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 648.321027] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.321154] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.321343] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 648.322256] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80a35f0d-2851-49f7-9029-7cc5b2b092b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.336153] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 648.336646] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 648.337832] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d3fcf78-7459-4815-9b3b-c27e4b840ed1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.346260] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 648.346260] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed3de4-f778-ed01-0e88-5325f8023180" [ 648.346260] env[65107]: _type = "Task" [ 648.346260] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.357985] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed3de4-f778-ed01-0e88-5325f8023180, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.401296] env[65107]: DEBUG nova.scheduler.client.report [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 648.410988] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.421391] env[65107]: DEBUG nova.compute.manager [req-f458791a-53f5-4c0c-84eb-fc04164e680e req-58935017-9871-43b5-bd4b-71140b1c1aa0 service nova] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Received event network-vif-deleted-b41e2bb0-3a13-473d-ac19-0e593444f482 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 648.423070] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] Releasing lock "refresh_cache-8306ff90-b8bd-4270-8133-96abe483156b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.423346] env[65107]: DEBUG nova.compute.manager [req-5d817504-59de-4861-a3f8-8ca4f75077db req-15ea8840-9f64-4d0c-9f58-24a620d14eda service nova] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Received event network-vif-deleted-74eb113e-24ab-43aa-bb1c-e5427ac59a2a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 648.485114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.485567] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.485863] env[65107]: DEBUG nova.network.neutron [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 648.511704] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Releasing lock "refresh_cache-ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.512607] env[65107]: DEBUG nova.compute.manager [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 648.512953] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.514429] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95255d45-3915-4d69-b5d2-f5079f3444c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.529852] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 648.530377] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4b00a1a-0398-407e-9b8b-abca0731f1b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.536156] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.537303] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.548331] env[65107]: DEBUG oslo_vmware.api [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 648.548331] env[65107]: value = "task-5102319" [ 648.548331] env[65107]: _type = "Task" [ 648.548331] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.566014] env[65107]: DEBUG oslo_vmware.api [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.588154] env[65107]: DEBUG nova.network.neutron [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 648.610261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.610261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.610261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.610534] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.610572] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.614561] env[65107]: INFO nova.compute.manager [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Terminating instance [ 648.648624] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.651062] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.756525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquiring lock "7062b119-01a4-45f4-bb52-212178382271" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.756525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "7062b119-01a4-45f4-bb52-212178382271" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.756525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "f7446d76-45c2-4e8b-981d-d37c230cf125" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.756525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.763891] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102318, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.866465] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed3de4-f778-ed01-0e88-5325f8023180, 'name': SearchDatastore_Task, 'duration_secs': 0.01745} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.867448] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16436410-e724-4ebe-a2ce-93e2157c8d8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.875106] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 648.875106] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523a1f14-a73d-748f-720a-8f291f30e12e" [ 648.875106] env[65107]: _type = "Task" [ 648.875106] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.886145] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523a1f14-a73d-748f-720a-8f291f30e12e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.914889] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.915470] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 648.920323] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.009s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.922092] env[65107]: INFO nova.compute.claims [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.989860] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.990708] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.991153] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.058907] env[65107]: DEBUG oslo_vmware.api [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102319, 'name': PowerOffVM_Task, 'duration_secs': 0.256339} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.059253] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 649.059422] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 649.059707] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc33f826-a5c4-4ec0-aace-7072513a3a5c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.090983] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 649.091177] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 649.091432] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleting the datastore file [datastore2] ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.091872] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39bfe31c-9afc-4bb4-ba53-83a8a44363f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.101146] env[65107]: DEBUG oslo_vmware.api [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for the task: (returnval){ [ 649.101146] env[65107]: value = "task-5102321" [ 649.101146] env[65107]: _type = "Task" [ 649.101146] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.111218] env[65107]: DEBUG oslo_vmware.api [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.121725] env[65107]: DEBUG nova.compute.manager [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 649.121940] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.122842] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da622be8-de84-4d60-87f0-9f90f02af15f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.131367] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 649.131760] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fcfe2203-edc0-4e10-8900-4308cf981bdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.139349] env[65107]: DEBUG oslo_vmware.api [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 649.139349] env[65107]: value = "task-5102322" [ 649.139349] env[65107]: _type = "Task" [ 649.139349] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.149614] env[65107]: DEBUG oslo_vmware.api [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.255188] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102318, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.264394] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 649.268611] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 649.389494] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523a1f14-a73d-748f-720a-8f291f30e12e, 'name': SearchDatastore_Task, 'duration_secs': 0.012948} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.389494] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.389921] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 1253473b-d4bc-49ac-9724-9aa1e7a8f038/1253473b-d4bc-49ac-9724-9aa1e7a8f038.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 649.390285] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67a88632-b95d-46de-ae18-3aad7742234c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.398764] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 649.398764] env[65107]: value = "task-5102323" [ 649.398764] env[65107]: _type = "Task" [ 649.398764] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.413949] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.428398] env[65107]: DEBUG nova.compute.utils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 649.435787] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 649.436051] env[65107]: DEBUG nova.network.neutron [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 649.436339] env[65107]: WARNING neutronclient.v2_0.client [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.436678] env[65107]: WARNING neutronclient.v2_0.client [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.437433] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.437775] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.612969] env[65107]: DEBUG oslo_vmware.api [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Task: {'id': task-5102321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.241652} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.614815] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 649.614815] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 649.614815] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.614815] env[65107]: INFO nova.compute.manager [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Took 1.10 seconds to destroy the instance on the hypervisor. [ 649.614815] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 649.615283] env[65107]: DEBUG nova.compute.manager [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 649.615283] env[65107]: DEBUG nova.network.neutron [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 649.615283] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.616157] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.616535] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.651080] env[65107]: DEBUG oslo_vmware.api [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102322, 'name': PowerOffVM_Task, 'duration_secs': 0.317802} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.652028] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 649.652028] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 649.652215] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41d1ae5b-86b5-4f16-9689-a5aa81eb28df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.736043] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 649.736325] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 649.736508] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleting the datastore file [datastore1] abb9dc0e-ce18-49c8-a472-d39a3fcec887 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.736794] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d96c8d66-0b60-4d8d-b3c8-c88bcec41216 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.745504] env[65107]: DEBUG oslo_vmware.api [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 649.745504] env[65107]: value = "task-5102325" [ 649.745504] env[65107]: _type = "Task" [ 649.745504] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.753394] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102318, 'name': Rename_Task, 'duration_secs': 1.296097} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.754186] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 649.754465] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e05259b0-fbd1-4013-b17a-e66c80ae0b0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.763190] env[65107]: DEBUG oslo_vmware.api [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102325, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.769343] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 649.769343] env[65107]: value = "task-5102326" [ 649.769343] env[65107]: _type = "Task" [ 649.769343] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.791646] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102326, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.794561] env[65107]: WARNING neutronclient.v2_0.client [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.795208] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.796093] env[65107]: WARNING openstack [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.805850] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.825575] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.890526] env[65107]: DEBUG nova.policy [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d169c6e1c1f4db8aaff217288b6ebc8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c5372cbba21404eb54eda5f08f0ac42', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 649.915111] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102323, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.957498] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 649.961098] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.962520] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.962520] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.962520] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.962520] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.964955] env[65107]: INFO nova.compute.manager [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Terminating instance [ 650.021492] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.021876] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.022139] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.022335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.022502] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.025325] env[65107]: INFO nova.compute.manager [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Terminating instance [ 650.052149] env[65107]: DEBUG nova.network.neutron [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 650.052482] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.140755] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.141338] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.211390] env[65107]: DEBUG nova.network.neutron [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Updating instance_info_cache with network_info: [{"id": "2275d82f-e805-4c7b-8656-ff9475e7f524", "address": "fa:16:3e:67:1e:59", "network": {"id": "c165d490-0569-4946-966c-3a5820469066", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1542494870-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "520207ef29e64aa799841fae36e469f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2275d82f-e8", "ovs_interfaceid": "2275d82f-e805-4c7b-8656-ff9475e7f524", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.257481] env[65107]: DEBUG oslo_vmware.api [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102325, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.457681} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.258109] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 650.258359] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 650.258593] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.258861] env[65107]: INFO nova.compute.manager [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Took 1.14 seconds to destroy the instance on the hypervisor. [ 650.259303] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 650.259779] env[65107]: DEBUG nova.compute.manager [-] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 650.259883] env[65107]: DEBUG nova.network.neutron [-] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 650.260178] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.261220] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.261220] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.281295] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102326, 'name': PowerOnVM_Task} progress is 92%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.412853] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.737786} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.413356] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 1253473b-d4bc-49ac-9724-9aa1e7a8f038/1253473b-d4bc-49ac-9724-9aa1e7a8f038.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 650.413617] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 650.414607] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0793987-f4fc-49ea-8ab0-45cf9944a060 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.423273] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 650.423273] env[65107]: value = "task-5102327" [ 650.423273] env[65107]: _type = "Task" [ 650.423273] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.433089] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.448463] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f48e06-162e-46c1-bf7f-79e19a9a9294 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.461755] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bb4a93-47ed-4e6c-8a66-aab6799016df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.475317] env[65107]: DEBUG nova.compute.manager [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 650.475494] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.478981] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83eecbe-0a99-4ebf-a8c6-d552ce232e69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.512631] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1695fe26-fa59-4973-be8e-d1d12237539b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.520611] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 650.521400] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a67d116a-f05a-4ff7-b9d6-1c5e54dc9762 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.527541] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de236cc8-ef61-416b-8640-07d790ea69cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.536573] env[65107]: DEBUG nova.compute.manager [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 650.536573] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.536573] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 650.536573] env[65107]: value = "task-5102328" [ 650.536573] env[65107]: _type = "Task" [ 650.536573] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.537458] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36e50a0-a3ab-41d2-8157-2e7a774ee7f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.554947] env[65107]: DEBUG nova.compute.provider_tree [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.559043] env[65107]: DEBUG nova.network.neutron [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.562453] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 650.562453] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c683771c-dbbc-46aa-b651-69b760e87c42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.566836] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.572935] env[65107]: DEBUG oslo_vmware.api [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 650.572935] env[65107]: value = "task-5102329" [ 650.572935] env[65107]: _type = "Task" [ 650.572935] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.584484] env[65107]: DEBUG oslo_vmware.api [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.586136] env[65107]: DEBUG nova.network.neutron [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Successfully created port: 8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 650.595708] env[65107]: WARNING neutronclient.v2_0.client [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.596404] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.596740] env[65107]: WARNING openstack [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.650335] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.715742] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Releasing lock "refresh_cache-e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.716503] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Instance network_info: |[{"id": "2275d82f-e805-4c7b-8656-ff9475e7f524", "address": "fa:16:3e:67:1e:59", "network": {"id": "c165d490-0569-4946-966c-3a5820469066", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1542494870-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "520207ef29e64aa799841fae36e469f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2275d82f-e8", "ovs_interfaceid": "2275d82f-e805-4c7b-8656-ff9475e7f524", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 650.716867] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:1e:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2275d82f-e805-4c7b-8656-ff9475e7f524', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 650.724704] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Creating folder: Project (520207ef29e64aa799841fae36e469f5). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 650.727389] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb2e4210-89cd-479d-ab8d-69bfa1faa138 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.742588] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Created folder: Project (520207ef29e64aa799841fae36e469f5) in parent group-v992574. [ 650.742795] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Creating folder: Instances. Parent ref: group-v992627. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 650.742878] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dcec7f6e-2052-4978-9650-386030dc27fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.755045] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Created folder: Instances in parent group-v992627. [ 650.755378] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 650.755596] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 650.755815] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6aede80-1a66-416d-9323-e45b89b26bd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.776469] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 650.776469] env[65107]: value = "task-5102332" [ 650.776469] env[65107]: _type = "Task" [ 650.776469] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.779806] env[65107]: DEBUG oslo_vmware.api [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102326, 'name': PowerOnVM_Task, 'duration_secs': 0.576849} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.783437] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 650.783754] env[65107]: INFO nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Took 10.68 seconds to spawn the instance on the hypervisor. [ 650.783833] env[65107]: DEBUG nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 650.784657] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488f9b73-ad15-4a76-bc2a-47085b983d4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.792731] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102332, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.937496] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113234} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.937946] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 650.939560] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aaf3675-14fb-4ba3-8fd9-da505b40483d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.980843] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 1253473b-d4bc-49ac-9724-9aa1e7a8f038/1253473b-d4bc-49ac-9724-9aa1e7a8f038.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 650.983332] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 650.986517] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-224b25da-378f-4d6d-8fe3-677cd1ad58a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.020093] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 651.020093] env[65107]: value = "task-5102333" [ 651.020093] env[65107]: _type = "Task" [ 651.020093] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.034843] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102333, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.051466] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 651.053048] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 651.053048] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 651.053048] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 651.053048] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 651.053509] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 651.053570] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.054039] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 651.054176] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 651.054487] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 651.054823] env[65107]: DEBUG nova.virt.hardware [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 651.056300] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d25b3d-bb2e-47b0-90d8-9dded62dd10b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.067746] env[65107]: DEBUG nova.scheduler.client.report [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 651.073796] env[65107]: INFO nova.compute.manager [-] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Took 1.46 seconds to deallocate network for instance. [ 651.074217] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.093107] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b55cec8-1bf8-4e89-b358-9fab5a9dba09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.110061] env[65107]: DEBUG oslo_vmware.api [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102329, 'name': PowerOffVM_Task, 'duration_secs': 0.409856} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.110237] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 651.110417] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 651.110733] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-267419d8-0ccb-4095-825a-6240ecbfc1a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.205092] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 651.205521] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 651.205636] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleting the datastore file [datastore2] 4e70aaf3-5f1c-4a61-a790-a0a5e409170d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 651.205926] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-413f9ef0-95e5-44d7-82f6-0ef3e0b519a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.214063] env[65107]: DEBUG oslo_vmware.api [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for the task: (returnval){ [ 651.214063] env[65107]: value = "task-5102335" [ 651.214063] env[65107]: _type = "Task" [ 651.214063] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.224983] env[65107]: DEBUG oslo_vmware.api [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102335, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.263638] env[65107]: DEBUG nova.network.neutron [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance_info_cache with network_info: [{"id": "086361c0-3cd7-4389-b5f0-c550a2adc834", "address": "fa:16:3e:bf:a0:a1", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086361c0-3c", "ovs_interfaceid": "086361c0-3cd7-4389-b5f0-c550a2adc834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 651.292788] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102332, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.307588] env[65107]: INFO nova.compute.manager [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Took 16.97 seconds to build instance. [ 651.532455] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.541060] env[65107]: DEBUG nova.network.neutron [-] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 651.553653] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.582841] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.663s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.583501] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 651.586318] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.539s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.587861] env[65107]: INFO nova.compute.claims [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.604034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.727754] env[65107]: DEBUG oslo_vmware.api [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Task: {'id': task-5102335, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176634} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.728041] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 651.728162] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 651.728329] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.728503] env[65107]: INFO nova.compute.manager [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Took 1.19 seconds to destroy the instance on the hypervisor. [ 651.728792] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 651.729134] env[65107]: DEBUG nova.compute.manager [-] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 651.729134] env[65107]: DEBUG nova.network.neutron [-] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 651.729394] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.730155] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.730622] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.769188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.793173] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102332, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.810775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86422c16-34f3-4f3b-92ad-8f8d3aa0cdfa tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "f58c3088-c821-4eec-be0a-81221debc98e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.484s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.032574] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102333, 'name': ReconfigVM_Task, 'duration_secs': 0.885331} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.033290] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 1253473b-d4bc-49ac-9724-9aa1e7a8f038/1253473b-d4bc-49ac-9724-9aa1e7a8f038.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 652.033611] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0e89e18-93c9-4afd-9671-514adcc83159 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.043299] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 652.043299] env[65107]: value = "task-5102336" [ 652.043299] env[65107]: _type = "Task" [ 652.043299] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.051012] env[65107]: INFO nova.compute.manager [-] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Took 1.79 seconds to deallocate network for instance. [ 652.061895] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102328, 'name': PowerOffVM_Task, 'duration_secs': 1.311873} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.062166] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102336, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.062703] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 652.062944] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 652.063173] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df30f15b-fa7e-4847-a3ce-7a01f3cc9dca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.093279] env[65107]: DEBUG nova.compute.utils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 652.097419] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 652.097687] env[65107]: DEBUG nova.network.neutron [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 652.098107] env[65107]: WARNING neutronclient.v2_0.client [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.098482] env[65107]: WARNING neutronclient.v2_0.client [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.099096] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.099451] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.143997] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 652.144245] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 652.145024] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleting the datastore file [datastore2] 0691bdfe-ff23-471b-ad94-2a98adcecd4f {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 652.145024] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dad78e20-72bb-4ed7-943a-88a0295f0ab6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.152480] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 652.152480] env[65107]: value = "task-5102338" [ 652.152480] env[65107]: _type = "Task" [ 652.152480] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.163573] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.294466] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102332, 'name': CreateVM_Task, 'duration_secs': 1.187302} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.294920] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 652.295191] env[65107]: WARNING neutronclient.v2_0.client [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.295616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.295752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.296106] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 652.296563] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ba057b8-1a6a-4a6d-a94f-1e57891ac3c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.302324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b2ec87-0a0e-4728-89c5-e5bb8b6c9168 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.308371] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 652.308371] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520e4d23-d354-e507-9a42-12710667ac27" [ 652.308371] env[65107]: _type = "Task" [ 652.308371] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.332188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002a6a61-99fc-4a87-a61e-b9af19e12f93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.339761] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520e4d23-d354-e507-9a42-12710667ac27, 'name': SearchDatastore_Task, 'duration_secs': 0.0326} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.340854] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.341028] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 652.341215] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.341379] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.341598] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 652.342129] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7063cfd1-3253-48e5-ac8f-67da79f93956 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.348065] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 652.367470] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 652.367713] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 652.368594] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d99eb393-62bf-48c8-9505-4718f11a3137 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.375893] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 652.375893] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f3fc2a-dd0b-cc4a-6299-e825e371b601" [ 652.375893] env[65107]: _type = "Task" [ 652.375893] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.386824] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f3fc2a-dd0b-cc4a-6299-e825e371b601, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.481658] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.557920] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102336, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.568757] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 652.612477] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 652.672193] env[65107]: DEBUG oslo_vmware.api [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.31058} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.672193] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 652.672193] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 652.672193] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.672193] env[65107]: INFO nova.compute.manager [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Took 2.20 seconds to destroy the instance on the hypervisor. [ 652.672843] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 652.672843] env[65107]: DEBUG nova.compute.manager [-] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 652.672843] env[65107]: DEBUG nova.network.neutron [-] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 652.673093] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.673876] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.674349] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.689427] env[65107]: DEBUG nova.network.neutron [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Successfully updated port: 8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 652.856278] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 652.856527] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9cd0da5-209b-4ec2-90c0-36cdf6c8d120 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.867663] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 652.867663] env[65107]: value = "task-5102339" [ 652.867663] env[65107]: _type = "Task" [ 652.867663] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.876832] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102339, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.892650] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f3fc2a-dd0b-cc4a-6299-e825e371b601, 'name': SearchDatastore_Task, 'duration_secs': 0.03255} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.893665] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37c54a38-cec8-4f70-a74e-648d0442c686 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.904373] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 652.904373] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5250d776-ed8a-79a3-9a75-42678a96a917" [ 652.904373] env[65107]: _type = "Task" [ 652.904373] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.913794] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5250d776-ed8a-79a3-9a75-42678a96a917, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.038639] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8cf9b5-2379-4a39-b61e-ecf39400e378 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.061507] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c27c83-ee12-45d2-993a-a2334441ad04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.069622] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102336, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.100398] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eccc110-583b-4a40-9120-c6a5527d58df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.109976] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0401ce-d0b2-4332-b66a-6cba7c438b0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.131053] env[65107]: DEBUG nova.compute.provider_tree [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.202848] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.203375] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquired lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.203931] env[65107]: DEBUG nova.network.neutron [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 653.382262] env[65107]: DEBUG oslo_vmware.api [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102339, 'name': PowerOnVM_Task, 'duration_secs': 0.516886} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.382662] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 653.383109] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8930e5f2-ff32-49f2-9eee-871d794696b8 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance 'fdb5603d-24b0-4fdd-b8cc-25c0ebee5311' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 653.420934] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5250d776-ed8a-79a3-9a75-42678a96a917, 'name': SearchDatastore_Task, 'duration_secs': 0.016484} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.422167] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.422569] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e40dc8ad-31cf-4798-ab97-3ee67c8fbe67/e40dc8ad-31cf-4798-ab97-3ee67c8fbe67.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 653.422976] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-843d65de-c542-4817-bb2c-986bccacdebb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.433673] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 653.433673] env[65107]: value = "task-5102340" [ 653.433673] env[65107]: _type = "Task" [ 653.433673] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.447584] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.557337] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102336, 'name': Rename_Task, 'duration_secs': 1.258865} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.558464] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 653.558792] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1629eb9c-be72-4dab-bf92-a44ae488ef11 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.566457] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 653.566457] env[65107]: value = "task-5102341" [ 653.566457] env[65107]: _type = "Task" [ 653.566457] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.579976] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102341, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.601636] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 653.632141] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 653.637145] env[65107]: DEBUG nova.scheduler.client.report [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 653.650244] env[65107]: DEBUG nova.policy [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '998c06c46eda4a84aa7ddd605721aaf3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aaf15dff8ae0472d94c9fecd0ba87a20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 653.667808] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 653.668294] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 653.668294] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 653.668398] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 653.668547] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 653.668695] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 653.668911] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.669142] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 653.669267] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 653.669430] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 653.669615] env[65107]: DEBUG nova.virt.hardware [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 653.670969] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e76d64e-cece-4126-9688-038c17a86a66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.680539] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc0d2d3-0bd9-4f79-a918-a83fce10398f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.707761] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.708184] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.945852] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102340, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.047833] env[65107]: DEBUG nova.network.neutron [-] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 654.079865] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.154310] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 654.154982] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 654.159096] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.003s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.159497] env[65107]: DEBUG nova.objects.instance [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lazy-loading 'resources' on Instance uuid eac4b8f9-9cd5-44a3-a12e-9ec22767b907 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 654.449628] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.952695} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.449944] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e40dc8ad-31cf-4798-ab97-3ee67c8fbe67/e40dc8ad-31cf-4798-ab97-3ee67c8fbe67.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 654.450131] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 654.450388] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-429080d1-8b52-4bdf-9785-50c6a64c55b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.465960] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 654.465960] env[65107]: value = "task-5102342" [ 654.465960] env[65107]: _type = "Task" [ 654.465960] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.484253] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.550722] env[65107]: INFO nova.compute.manager [-] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Took 2.82 seconds to deallocate network for instance. [ 654.582276] env[65107]: DEBUG oslo_vmware.api [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102341, 'name': PowerOnVM_Task, 'duration_secs': 0.738541} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.583073] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 654.583384] env[65107]: INFO nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Took 11.91 seconds to spawn the instance on the hypervisor. [ 654.583570] env[65107]: DEBUG nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 654.584475] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aaad1d1-4fec-4096-8a9f-42b599317da1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.668612] env[65107]: DEBUG nova.compute.utils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 654.674373] env[65107]: DEBUG nova.network.neutron [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 654.680789] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 654.683132] env[65107]: DEBUG nova.network.neutron [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 654.683819] env[65107]: WARNING neutronclient.v2_0.client [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.684826] env[65107]: WARNING neutronclient.v2_0.client [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.688089] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.689397] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.984653] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213735} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.984814] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 654.985971] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353bbf1c-d955-4a78-a191-8d74663c3d53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.992824] env[65107]: DEBUG nova.network.neutron [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Successfully created port: d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 655.026570] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] e40dc8ad-31cf-4798-ab97-3ee67c8fbe67/e40dc8ad-31cf-4798-ab97-3ee67c8fbe67.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 655.031290] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af39dc2d-eeb3-4096-ac85-556d2c286c15 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.052723] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 655.052723] env[65107]: value = "task-5102343" [ 655.052723] env[65107]: _type = "Task" [ 655.052723] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.058842] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.065432] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.098397] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Received event network-vif-plugged-321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 655.102777] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "f58c3088-c821-4eec-be0a-81221debc98e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.102777] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Lock "f58c3088-c821-4eec-be0a-81221debc98e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.102777] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Lock "f58c3088-c821-4eec-be0a-81221debc98e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.102777] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] No waiting events found dispatching network-vif-plugged-321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 655.102777] env[65107]: WARNING nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Received unexpected event network-vif-plugged-321da8b7-0ae1-438d-816f-6c711ef9b003 for instance with vm_state active and task_state None. [ 655.103082] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Received event network-changed-321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 655.103082] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Refreshing instance network info cache due to event network-changed-321da8b7-0ae1-438d-816f-6c711ef9b003. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 655.103082] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "refresh_cache-f58c3088-c821-4eec-be0a-81221debc98e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.103082] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquired lock "refresh_cache-f58c3088-c821-4eec-be0a-81221debc98e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.103082] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Refreshing network info cache for port 321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 655.114745] env[65107]: DEBUG nova.network.neutron [-] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 655.118035] env[65107]: INFO nova.compute.manager [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Took 19.28 seconds to build instance. [ 655.168206] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.168424] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.176474] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 655.198204] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c3d48e-3f6b-465a-81a9-a87ebd8b3460 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.208236] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fb50e9-e77f-4c3e-a296-279dcf67bdf0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.215430] env[65107]: DEBUG nova.policy [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '612b5e91ecf144d9ba4c331085d239f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9fcd71f4f064b488f832406fedafbe0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 655.990272] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.991836] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.991836] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.998328] env[65107]: INFO nova.compute.manager [-] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Took 3.33 seconds to deallocate network for instance. [ 656.000107] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73544a2b-217e-4439-ace0-b02bf75a56fa tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.172s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.008275] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2100f1e-9ab7-472e-bc2e-35fd995281d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.023643] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102343, 'name': ReconfigVM_Task, 'duration_secs': 0.557052} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.024044] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Reconfigured VM instance instance-00000013 to attach disk [datastore1] e40dc8ad-31cf-4798-ab97-3ee67c8fbe67/e40dc8ad-31cf-4798-ab97-3ee67c8fbe67.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 656.025712] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ccd1f3-62ec-4d5b-8964-2ef75cc4de72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.029939] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a16c191-f9ed-4228-a4d1-9ca94954c8ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.044511] env[65107]: DEBUG nova.compute.provider_tree [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.050020] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 656.050020] env[65107]: value = "task-5102344" [ 656.050020] env[65107]: _type = "Task" [ 656.050020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.058141] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102344, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.505271] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 656.516440] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.534780] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 656.535108] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 656.535278] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 656.535490] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 656.535969] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 656.535969] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 656.535969] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.536165] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 656.536873] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 656.536873] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 656.536873] env[65107]: DEBUG nova.virt.hardware [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 656.537496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9111555c-fe18-4737-974c-773c6b721be9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.545591] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5a24b8-3cf9-4094-b8a7-59933484282c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.550103] env[65107]: DEBUG nova.scheduler.client.report [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.561876] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102344, 'name': Rename_Task, 'duration_secs': 0.444547} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.572731] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 656.574010] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31c97c52-0599-4389-922d-59b405afd632 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.581056] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 656.581056] env[65107]: value = "task-5102345" [ 656.581056] env[65107]: _type = "Task" [ 656.581056] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.589932] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.881858] env[65107]: DEBUG nova.network.neutron [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Successfully updated port: d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 657.028492] env[65107]: WARNING neutronclient.v2_0.client [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.029233] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.029666] env[65107]: WARNING openstack [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.057981] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.899s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.060407] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.255s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.063257] env[65107]: INFO nova.compute.claims [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.095772] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.097815] env[65107]: INFO nova.scheduler.client.report [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted allocations for instance eac4b8f9-9cd5-44a3-a12e-9ec22767b907 [ 657.385081] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "refresh_cache-7b24c530-56bf-4666-96f9-e2dc66728f1a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.385328] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired lock "refresh_cache-7b24c530-56bf-4666-96f9-e2dc66728f1a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.385698] env[65107]: DEBUG nova.network.neutron [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 657.395800] env[65107]: DEBUG nova.network.neutron [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Successfully created port: 55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 657.480475] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.480954] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.586249] env[65107]: DEBUG nova.network.neutron [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Updating instance_info_cache with network_info: [{"id": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "address": "fa:16:3e:92:8d:9e", "network": {"id": "b9e7c478-0189-4385-98a5-bc3ca80526e5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1623056907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c5372cbba21404eb54eda5f08f0ac42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea6e81c3-94aa-40a6-a4d4-7f338b503442", "external-id": "nsx-vlan-transportzone-637", "segmentation_id": 637, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8226470b-4e", "ovs_interfaceid": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 657.602801] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.613025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8753e28b-c25b-409e-b693-180ca8e79fcd tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "eac4b8f9-9cd5-44a3-a12e-9ec22767b907" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.922s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.892029] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.892029] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.901397] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.902289] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.902810] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.091107] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Releasing lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.091598] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Instance network_info: |[{"id": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "address": "fa:16:3e:92:8d:9e", "network": {"id": "b9e7c478-0189-4385-98a5-bc3ca80526e5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1623056907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c5372cbba21404eb54eda5f08f0ac42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea6e81c3-94aa-40a6-a4d4-7f338b503442", "external-id": "nsx-vlan-transportzone-637", "segmentation_id": 637, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8226470b-4e", "ovs_interfaceid": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 658.095622] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:8d:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea6e81c3-94aa-40a6-a4d4-7f338b503442', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8226470b-4e4b-4b3f-b40f-b29d3149af69', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 658.103532] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Creating folder: Project (9c5372cbba21404eb54eda5f08f0ac42). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 658.107034] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-802692e6-002c-4a50-9eab-7eec4364ea52 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.109987] env[65107]: DEBUG nova.network.neutron [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 658.118777] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.122815] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Created folder: Project (9c5372cbba21404eb54eda5f08f0ac42) in parent group-v992574. [ 658.122815] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Creating folder: Instances. Parent ref: group-v992630. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 658.122815] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c69f2e8a-e996-40eb-9b32-fb9e236b7a68 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.136161] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Created folder: Instances in parent group-v992630. [ 658.136489] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 658.136726] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 658.137268] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d4a2271-0465-4f15-b167-b4c634ae05ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.165190] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 658.165190] env[65107]: value = "task-5102348" [ 658.165190] env[65107]: _type = "Task" [ 658.165190] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.176736] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102348, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.268161] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.268288] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.359649] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Updated VIF entry in instance network info cache for port 321da8b7-0ae1-438d-816f-6c711ef9b003. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 658.360287] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Updating instance_info_cache with network_info: [{"id": "321da8b7-0ae1-438d-816f-6c711ef9b003", "address": "fa:16:3e:b6:63:49", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap321da8b7-0a", "ovs_interfaceid": "321da8b7-0ae1-438d-816f-6c711ef9b003", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 658.501083] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e84741-751d-47dd-a44c-abcb9f84081d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.511863] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af2077d-f2e3-4f9f-8c83-15b1de7c7978 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.550430] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e22b5b3-45e1-4916-bbdf-ae67d0d4b580 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.562222] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5e1f6e-a9a2-4be4-8f7c-f5248d469cdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.576944] env[65107]: DEBUG nova.compute.provider_tree [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.601737] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.681286] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102348, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.797706] env[65107]: DEBUG nova.compute.manager [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Received event network-changed-3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 658.798020] env[65107]: DEBUG nova.compute.manager [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Refreshing instance network info cache due to event network-changed-3bb17344-0031-4226-9117-1366ecf93330. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 658.798192] env[65107]: DEBUG oslo_concurrency.lockutils [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Acquiring lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.798349] env[65107]: DEBUG oslo_concurrency.lockutils [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Acquired lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.798509] env[65107]: DEBUG nova.network.neutron [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Refreshing network info cache for port 3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 658.863481] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Releasing lock "refresh_cache-f58c3088-c821-4eec-be0a-81221debc98e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.863765] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received event network-vif-plugged-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 658.864033] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 658.865229] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 658.865522] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 658.866923] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] No waiting events found dispatching network-vif-plugged-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 658.867209] env[65107]: WARNING nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received unexpected event network-vif-plugged-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e for instance with vm_state active and task_state None. [ 658.867411] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 658.867575] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing instance network info cache due to event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 658.867827] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.867923] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.868233] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 658.981982] env[65107]: WARNING neutronclient.v2_0.client [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.982676] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.983056] env[65107]: WARNING openstack [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.080562] env[65107]: DEBUG nova.scheduler.client.report [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.101905] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.132714] env[65107]: DEBUG nova.network.neutron [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Successfully updated port: 55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 659.175997] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102348, 'name': CreateVM_Task, 'duration_secs': 0.517566} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.180584] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 659.181532] env[65107]: WARNING neutronclient.v2_0.client [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.181908] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.182072] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.182385] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 659.182952] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-703b73ca-be31-449c-9db4-2fc3e78ac690 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.188185] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 659.188185] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52370033-9984-9fc8-552a-213fd0de8fc4" [ 659.188185] env[65107]: _type = "Task" [ 659.188185] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.200241] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52370033-9984-9fc8-552a-213fd0de8fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.262264] env[65107]: DEBUG nova.network.neutron [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Updating instance_info_cache with network_info: [{"id": "d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff", "address": "fa:16:3e:43:32:e9", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd533dd9c-4b", "ovs_interfaceid": "d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.301865] env[65107]: WARNING neutronclient.v2_0.client [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.302700] env[65107]: WARNING openstack [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.303202] env[65107]: WARNING openstack [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.371729] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.372463] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.372838] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.585834] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.586272] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 659.589978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.765s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.591690] env[65107]: INFO nova.compute.claims [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.606649] env[65107]: DEBUG oslo_vmware.api [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102345, 'name': PowerOnVM_Task, 'duration_secs': 2.662657} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.607245] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 659.607613] env[65107]: INFO nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Took 13.33 seconds to spawn the instance on the hypervisor. [ 659.608140] env[65107]: DEBUG nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 659.610097] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbf3c1f-032c-44bd-85eb-863b4c65bb67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.637409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.637409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquired lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.637409] env[65107]: DEBUG nova.network.neutron [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 659.665361] env[65107]: WARNING openstack [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.665361] env[65107]: WARNING openstack [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.684868] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.685284] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.705348] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52370033-9984-9fc8-552a-213fd0de8fc4, 'name': SearchDatastore_Task, 'duration_secs': 0.019864} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.705701] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.705931] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.706360] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.706360] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.706555] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 659.707165] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1330f194-b981-4518-9024-9ca2e7d0b0a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.722670] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 659.722868] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 659.723685] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6f91a65-6f7a-4739-b8ba-e75264f6bd45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.732029] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 659.732029] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b7859-e418-d37e-b434-08c0af97a17f" [ 659.732029] env[65107]: _type = "Task" [ 659.732029] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.739625] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b7859-e418-d37e-b434-08c0af97a17f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.765031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Releasing lock "refresh_cache-7b24c530-56bf-4666-96f9-e2dc66728f1a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.765031] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Instance network_info: |[{"id": "d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff", "address": "fa:16:3e:43:32:e9", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd533dd9c-4b", "ovs_interfaceid": "d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 659.765296] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:32:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd533dd9c-4bb6-4f96-b6fb-84bd5623a8ff', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 659.775136] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 659.776030] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 659.776402] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c82da10-d80d-4e21-8572-23442bdc8c6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.795794] env[65107]: WARNING neutronclient.v2_0.client [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.796613] env[65107]: WARNING openstack [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.797122] env[65107]: WARNING openstack [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.806933] env[65107]: DEBUG nova.compute.manager [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 659.808823] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d6672c-e2ef-4d17-9814-bca322953057 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.814286] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.815214] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.815710] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.826272] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 659.826272] env[65107]: value = "task-5102349" [ 659.826272] env[65107]: _type = "Task" [ 659.826272] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.845248] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102349, 'name': CreateVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.092499] env[65107]: DEBUG nova.compute.utils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 660.094793] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 660.095289] env[65107]: DEBUG nova.network.neutron [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 660.095618] env[65107]: WARNING neutronclient.v2_0.client [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.096268] env[65107]: WARNING neutronclient.v2_0.client [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.096946] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.097852] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.139970] env[65107]: INFO nova.compute.manager [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Took 23.58 seconds to build instance. [ 660.142184] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.142184] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.242291] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b7859-e418-d37e-b434-08c0af97a17f, 'name': SearchDatastore_Task, 'duration_secs': 0.047749} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.243278] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cccea16-759e-4a83-bcc5-2c4cff3f0645 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.251095] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 660.251095] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eb8c37-2587-81f1-9370-26e82bb6df3c" [ 660.251095] env[65107]: _type = "Task" [ 660.251095] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.261847] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eb8c37-2587-81f1-9370-26e82bb6df3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.263513] env[65107]: DEBUG nova.network.neutron [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Updated VIF entry in instance network info cache for port 3bb17344-0031-4226-9117-1366ecf93330. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 660.263966] env[65107]: DEBUG nova.network.neutron [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Updating instance_info_cache with network_info: [{"id": "3bb17344-0031-4226-9117-1366ecf93330", "address": "fa:16:3e:06:8b:e4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bb17344-00", "ovs_interfaceid": "3bb17344-0031-4226-9117-1366ecf93330", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.304581] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updated VIF entry in instance network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 660.304581] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.340386] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102349, 'name': CreateVM_Task, 'duration_secs': 0.413526} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.341659] env[65107]: INFO nova.compute.manager [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] instance snapshotting [ 660.343381] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 660.344274] env[65107]: WARNING neutronclient.v2_0.client [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.344650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.344800] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.345142] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 660.345974] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b81c36-5cf0-4be5-8b0c-13ff763d2b8d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.348853] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e142885-38c2-4c06-b422-0ccbe4bcce56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.371554] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76280ff5-69d4-4f28-b0bc-a61865edddab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.374694] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 660.374694] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ade50e-06cf-80f3-5e01-ee730f2042d0" [ 660.374694] env[65107]: _type = "Task" [ 660.374694] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.388470] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ade50e-06cf-80f3-5e01-ee730f2042d0, 'name': SearchDatastore_Task, 'duration_secs': 0.014568} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.388761] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.388992] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 660.389220] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.412166] env[65107]: DEBUG nova.compute.manager [req-7db8d8e7-ad51-4336-9e7a-c6b3bff6d9da req-e8c1d437-6aa4-419b-9ce8-b0c249eb2d27 service nova] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Received event network-vif-deleted-d06bc826-4580-47a4-a3b4-841380ee155f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 660.412480] env[65107]: DEBUG nova.compute.manager [req-7db8d8e7-ad51-4336-9e7a-c6b3bff6d9da req-e8c1d437-6aa4-419b-9ce8-b0c249eb2d27 service nova] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Received event network-vif-deleted-d2435e56-5f0a-44e9-8154-2f2dd4103c55 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 660.412551] env[65107]: DEBUG nova.compute.manager [req-7db8d8e7-ad51-4336-9e7a-c6b3bff6d9da req-e8c1d437-6aa4-419b-9ce8-b0c249eb2d27 service nova] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Received event network-vif-deleted-d9b6f78e-710c-4b4c-a9bd-04632f29764a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 660.610140] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 660.653318] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa876657-77ac-43e2-a80d-c4e062c715b6 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.109s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 660.762068] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eb8c37-2587-81f1-9370-26e82bb6df3c, 'name': SearchDatastore_Task, 'duration_secs': 0.012605} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.762350] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.762607] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 705b91ea-ccd2-41ce-a14f-2ea5b70f80df/705b91ea-ccd2-41ce-a14f-2ea5b70f80df.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 660.762886] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.763147] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 660.763377] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c2faf07-9f18-479f-bdf3-e193dcf9025a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.769859] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd1eb54c-e03f-4717-8cd2-1403c2d16095 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.770941] env[65107]: DEBUG oslo_concurrency.lockutils [req-76044f87-c227-4043-adbe-0aefccb4edae req-561731fc-28d4-4616-b7c6-12b7f0862d82 service nova] Releasing lock "refresh_cache-be518c1d-edd7-40ff-b7cc-3310885b07b4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.778570] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 660.778570] env[65107]: value = "task-5102350" [ 660.778570] env[65107]: _type = "Task" [ 660.778570] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.779927] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 660.780218] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 660.784891] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51bd309d-d971-4439-8c77-5be8cefadb04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.795631] env[65107]: DEBUG nova.policy [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e653221cd6da4b4780fec3fb398d90cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3fa2b26480f4495926fadf02489de3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 660.800640] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 660.800640] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525a763e-34cb-7769-e196-80f8e935a118" [ 660.800640] env[65107]: _type = "Task" [ 660.800640] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.809033] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.813760] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.814222] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Received event network-changed-02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 660.814534] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Refreshing instance network info cache due to event network-changed-02108638-8e1e-447a-9b94-a45adfd32161. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 660.814861] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.815246] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquired lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.815565] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Refreshing network info cache for port 02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 660.823313] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525a763e-34cb-7769-e196-80f8e935a118, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.842165] env[65107]: DEBUG nova.network.neutron [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 660.884276] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 660.884886] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-77b26947-1800-4842-948a-9db07262ea23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.896084] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 660.896084] env[65107]: value = "task-5102351" [ 660.896084] env[65107]: _type = "Task" [ 660.896084] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.910590] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102351, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.032584] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90c6747-6b3c-41b9-9343-fe15850421d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.042593] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc635434-3463-4026-ade8-14e7bcff01ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.082099] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.082099] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.088829] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f25b52-7248-4279-bd47-4117801bbb83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.097732] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebd9eb6-854a-4583-8601-cafce9f08b41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.112614] env[65107]: DEBUG nova.compute.provider_tree [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.117334] env[65107]: INFO nova.virt.block_device [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Booting with volume 7ab345f6-458d-4bb0-be22-c593a481ce3f at /dev/sda [ 661.164103] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d053cc8-cf36-4ab0-811c-dd1b33037b95 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.174940] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84e91e1-9e2d-4f7e-9c96-db7f886cc5f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.218905] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34819425-e627-4498-873c-16167b2867bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.229669] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdec751-2c04-4a7a-9f55-2c869a2307af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.270200] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0374b651-29e3-48d7-aac3-fa2750f65065 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.375968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2616cb27-a99d-49c6-a2cb-a89d7c741b9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.375968] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102350, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.375968] env[65107]: DEBUG nova.virt.block_device [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updating existing volume attachment record: b97e3c6b-a27d-4c54-a315-be3047baf127 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 661.375968] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.375968] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.377692] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.377692] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525a763e-34cb-7769-e196-80f8e935a118, 'name': SearchDatastore_Task, 'duration_secs': 0.030403} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.377692] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd8fc706-8573-47d1-8788-bf6ce4b93c34 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.377692] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 661.377692] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5217a5d7-7461-a333-3837-91d1fa853554" [ 661.377692] env[65107]: _type = "Task" [ 661.377692] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.378951] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5217a5d7-7461-a333-3837-91d1fa853554, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.411136] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102351, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.616019] env[65107]: DEBUG nova.scheduler.client.report [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.796447] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102350, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.746841} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.796980] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 705b91ea-ccd2-41ce-a14f-2ea5b70f80df/705b91ea-ccd2-41ce-a14f-2ea5b70f80df.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 661.797326] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 661.797731] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f37fab10-fb21-456a-977f-36b9dde6715c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.807740] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 661.807740] env[65107]: value = "task-5102352" [ 661.807740] env[65107]: _type = "Task" [ 661.807740] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.817142] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102352, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.853760] env[65107]: WARNING neutronclient.v2_0.client [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.854535] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.855089] env[65107]: WARNING openstack [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.875289] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5217a5d7-7461-a333-3837-91d1fa853554, 'name': SearchDatastore_Task, 'duration_secs': 0.069368} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.875289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.875289] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 7b24c530-56bf-4666-96f9-e2dc66728f1a/7b24c530-56bf-4666-96f9-e2dc66728f1a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 661.875289] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a6d6f4a-5ae5-410d-ab04-254ab791a2e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.884816] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 661.884816] env[65107]: value = "task-5102353" [ 661.884816] env[65107]: _type = "Task" [ 661.884816] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.898697] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.911511] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102351, 'name': CreateSnapshot_Task, 'duration_secs': 0.881789} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.911511] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 661.911511] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb83034-0ea4-4a4c-858b-13ac3abbc7f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.925465] env[65107]: DEBUG nova.network.neutron [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Successfully created port: 09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 662.127250] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.129018] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 662.132570] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.531s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.132948] env[65107]: DEBUG nova.objects.instance [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lazy-loading 'resources' on Instance uuid ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 662.163160] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.163160] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.282212] env[65107]: DEBUG nova.network.neutron [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Updating instance_info_cache with network_info: [{"id": "55102692-2502-4633-b00b-cc29760fb310", "address": "fa:16:3e:6a:99:59", "network": {"id": "90ceb5af-809c-4b48-9c35-26e48235753c", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-156477542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e9fcd71f4f064b488f832406fedafbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55102692-25", "ovs_interfaceid": "55102692-2502-4633-b00b-cc29760fb310", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 662.325275] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102352, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126216} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.325593] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 662.326557] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7667b4b3-b65f-4354-bb1f-a46518ea58e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.360396] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] 705b91ea-ccd2-41ce-a14f-2ea5b70f80df/705b91ea-ccd2-41ce-a14f-2ea5b70f80df.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 662.366161] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daed3727-3834-48cc-8eb1-c08876944f07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.407100] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102353, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.413616] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 662.413616] env[65107]: value = "task-5102354" [ 662.413616] env[65107]: _type = "Task" [ 662.413616] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.437077] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 662.437740] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cdc337f2-c505-4b26-a9fe-b5ab4126f6bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.453855] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 662.453855] env[65107]: value = "task-5102355" [ 662.453855] env[65107]: _type = "Task" [ 662.453855] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.467928] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102355, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.507337] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.508139] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.508533] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.637555] env[65107]: DEBUG nova.compute.utils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 662.642458] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 662.642692] env[65107]: DEBUG nova.network.neutron [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 662.643208] env[65107]: WARNING neutronclient.v2_0.client [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.643450] env[65107]: WARNING neutronclient.v2_0.client [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.644044] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.645028] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.763500] env[65107]: DEBUG nova.policy [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '145e74601afc4e99860f5264b20c6dfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95393e0b56bd47cd8bb37207e0500269', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 662.787058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Releasing lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 662.787329] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Instance network_info: |[{"id": "55102692-2502-4633-b00b-cc29760fb310", "address": "fa:16:3e:6a:99:59", "network": {"id": "90ceb5af-809c-4b48-9c35-26e48235753c", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-156477542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e9fcd71f4f064b488f832406fedafbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55102692-25", "ovs_interfaceid": "55102692-2502-4633-b00b-cc29760fb310", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 662.791182] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:99:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e350f83a-f581-4e10-ac16-0b0f7bfd3d38', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '55102692-2502-4633-b00b-cc29760fb310', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 662.799946] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Creating folder: Project (e9fcd71f4f064b488f832406fedafbe0). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 662.800962] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a56922ee-c452-4c69-84f6-c0e233c19899 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.818287] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Created folder: Project (e9fcd71f4f064b488f832406fedafbe0) in parent group-v992574. [ 662.818456] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Creating folder: Instances. Parent ref: group-v992636. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 662.818693] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-768c9fd5-3b86-43e2-b5a4-7158e86f73c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.828877] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updated VIF entry in instance network info cache for port 02108638-8e1e-447a-9b94-a45adfd32161. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 662.828877] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updating instance_info_cache with network_info: [{"id": "02108638-8e1e-447a-9b94-a45adfd32161", "address": "fa:16:3e:49:92:f3", "network": {"id": "f011d711-ac42-4e0b-bcd5-8f569242741c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-475912691-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac04479a6164cf7b18b9ad304a0c9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02108638-8e", "ovs_interfaceid": "02108638-8e1e-447a-9b94-a45adfd32161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 662.839862] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Created folder: Instances in parent group-v992636. [ 662.840564] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 662.841374] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 662.841549] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1098ed69-59b8-41b8-ac0d-4a8994bc98ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.872470] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 662.872470] env[65107]: value = "task-5102358" [ 662.872470] env[65107]: _type = "Task" [ 662.872470] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.882286] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102358, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.899246] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.72393} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.899710] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 7b24c530-56bf-4666-96f9-e2dc66728f1a/7b24c530-56bf-4666-96f9-e2dc66728f1a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 662.899929] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 662.900291] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-738af0ae-7982-4057-956b-82d5f719c4d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.918385] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 662.918385] env[65107]: value = "task-5102359" [ 662.918385] env[65107]: _type = "Task" [ 662.918385] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.931350] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "ad53146e-210d-4321-b819-90795c1af260" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 662.931751] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "ad53146e-210d-4321-b819-90795c1af260" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.941116] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102354, 'name': ReconfigVM_Task, 'duration_secs': 0.388204} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.947420] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Reconfigured VM instance instance-00000014 to attach disk [datastore1] 705b91ea-ccd2-41ce-a14f-2ea5b70f80df/705b91ea-ccd2-41ce-a14f-2ea5b70f80df.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 662.947420] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55a42807-fc4f-4796-b57c-6d6f429007aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.948657] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102359, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.955718] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 662.955718] env[65107]: value = "task-5102360" [ 662.955718] env[65107]: _type = "Task" [ 662.955718] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.980550] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102360, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.981060] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102355, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.121531] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7079e674-078c-4e2a-b656-e330c1696cc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.132327] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de2a198-10c3-4b61-bbe5-11a51bf59424 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.165352] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 663.168784] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7280a0-e539-4c35-873f-6c70f4d88ee1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.179509] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50552376-9510-4b18-9064-827f7cd6604a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.198496] env[65107]: DEBUG nova.compute.provider_tree [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.296622] env[65107]: DEBUG nova.network.neutron [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Successfully created port: 8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 663.330445] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.330857] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.331185] env[65107]: DEBUG nova.compute.manager [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Going to confirm migration 1 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 663.333298] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Releasing lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.333529] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Received event network-vif-plugged-2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 663.333738] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.333939] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.334257] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 663.334502] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] No waiting events found dispatching network-vif-plugged-2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 663.334703] env[65107]: WARNING nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Received unexpected event network-vif-plugged-2275d82f-e805-4c7b-8656-ff9475e7f524 for instance with vm_state building and task_state spawning. [ 663.334878] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Received event network-changed-2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 663.335105] env[65107]: DEBUG nova.compute.manager [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Refreshing instance network info cache due to event network-changed-2275d82f-e805-4c7b-8656-ff9475e7f524. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 663.335360] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquiring lock "refresh_cache-e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.335535] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Acquired lock "refresh_cache-e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.335711] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Refreshing network info cache for port 2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 663.381915] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102358, 'name': CreateVM_Task, 'duration_secs': 0.453202} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.382085] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 663.382798] env[65107]: WARNING neutronclient.v2_0.client [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.383362] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.383568] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.383933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 663.384579] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74a2dc94-3e2f-46ce-92d4-b2c750adb327 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.396403] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "a2b72bac-6806-4b7b-b972-32aea52f7c82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.396705] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "a2b72bac-6806-4b7b-b972-32aea52f7c82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.397932] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 663.397932] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d9c7d6-c12f-9ab4-7403-4ca47b2b16fe" [ 663.397932] env[65107]: _type = "Task" [ 663.397932] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.408915] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d9c7d6-c12f-9ab4-7403-4ca47b2b16fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.431853] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102359, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079224} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.432269] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 663.433998] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f33b8a-895e-4403-a34c-f841b4a5389f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.452149] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 663.464572] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] 7b24c530-56bf-4666-96f9-e2dc66728f1a/7b24c530-56bf-4666-96f9-e2dc66728f1a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 663.465481] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8b6eca0-dc0c-4d10-b213-7de50ed689dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.488116] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 663.488815] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 663.489092] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 663.489279] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 663.489467] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 663.489618] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 663.489774] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 663.490032] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.490206] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 663.490397] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 663.490584] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 663.490773] env[65107]: DEBUG nova.virt.hardware [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 663.492393] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759fbaab-159e-4e7d-ad8c-0efbc8b37899 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.497826] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 663.497826] env[65107]: value = "task-5102361" [ 663.497826] env[65107]: _type = "Task" [ 663.497826] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.501971] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102355, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.508056] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102360, 'name': Rename_Task, 'duration_secs': 0.192849} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.510930] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 663.512024] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7443da30-8e9f-4d67-8650-c256ed48935e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.514871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c68df4-c67c-4af9-8b9e-3da746576c4d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.522846] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.535946] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 663.535946] env[65107]: value = "task-5102362" [ 663.535946] env[65107]: _type = "Task" [ 663.535946] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.545836] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102362, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.703767] env[65107]: DEBUG nova.scheduler.client.report [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.748816] env[65107]: DEBUG nova.network.neutron [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Successfully updated port: 09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 663.841259] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.842077] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.842515] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.850778] env[65107]: WARNING neutronclient.v2_0.client [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.899774] env[65107]: DEBUG nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 663.923387] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d9c7d6-c12f-9ab4-7403-4ca47b2b16fe, 'name': SearchDatastore_Task, 'duration_secs': 0.041611} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.923804] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.924870] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.925227] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.925402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.925984] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.926470] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c814a27-d577-40c5-9904-cc5e650f16a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.937726] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.937983] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 663.938876] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12835cb9-e1af-4b1f-ad13-eb8de377caea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.947090] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 663.947090] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ea7e78-7e2d-ce71-0b23-059b1cc1df73" [ 663.947090] env[65107]: _type = "Task" [ 663.947090] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.960809] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ea7e78-7e2d-ce71-0b23-059b1cc1df73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.986172] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102355, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.996900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.013852] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102361, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.046643] env[65107]: DEBUG oslo_vmware.api [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102362, 'name': PowerOnVM_Task, 'duration_secs': 0.487731} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.046972] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 664.047201] env[65107]: INFO nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Took 13.06 seconds to spawn the instance on the hypervisor. [ 664.047381] env[65107]: DEBUG nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 664.048218] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb478e6d-966b-424a-85d1-5228cbb762fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.178964] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 664.210128] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 664.210128] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 664.210128] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 664.210365] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 664.210365] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 664.210365] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 664.210365] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.210365] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 664.210540] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 664.210540] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 664.210540] env[65107]: DEBUG nova.virt.hardware [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 664.211463] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.079s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.214307] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b973288-d77c-4360-a51a-c880f91cff49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.217909] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.652s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.217909] env[65107]: DEBUG nova.objects.instance [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lazy-loading 'resources' on Instance uuid abb9dc0e-ce18-49c8-a472-d39a3fcec887 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 664.225487] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c0a1fe-16b3-4f78-b976-6bee2e4e931f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.245332] env[65107]: INFO nova.scheduler.client.report [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Deleted allocations for instance ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1 [ 664.253602] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquiring lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.253910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquired lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 664.254148] env[65107]: DEBUG nova.network.neutron [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 664.439119] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.459824] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ea7e78-7e2d-ce71-0b23-059b1cc1df73, 'name': SearchDatastore_Task, 'duration_secs': 0.016828} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.461063] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d537800-a25e-420e-befe-db0cb9e849ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.467273] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 664.467273] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52243b59-f11b-b690-783b-f23c0d9d2029" [ 664.467273] env[65107]: _type = "Task" [ 664.467273] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.479944] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102355, 'name': CloneVM_Task, 'duration_secs': 1.929263} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.483032] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Created linked-clone VM from snapshot [ 664.483361] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52243b59-f11b-b690-783b-f23c0d9d2029, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.483896] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a69be2f-9a65-4b8a-94d0-c78ecace676a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.491841] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Uploading image a656a63b-c667-4970-863b-5666f3f5df1d {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 664.516484] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102361, 'name': ReconfigVM_Task, 'duration_secs': 0.57191} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.516786] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Reconfigured VM instance instance-00000015 to attach disk [datastore1] 7b24c530-56bf-4666-96f9-e2dc66728f1a/7b24c530-56bf-4666-96f9-e2dc66728f1a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 664.522344] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 664.522344] env[65107]: value = "vm-992635" [ 664.522344] env[65107]: _type = "VirtualMachine" [ 664.522344] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 664.522606] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c49da777-7c8c-4e6b-8c4b-53f3f4278a50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.524636] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8ac72c38-a649-4ea4-84e4-fd16dc59f96a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.533663] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 664.533663] env[65107]: value = "task-5102363" [ 664.533663] env[65107]: _type = "Task" [ 664.533663] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.535170] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lease: (returnval){ [ 664.535170] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f9598-fbb8-29f4-8c58-4119fe5719f8" [ 664.535170] env[65107]: _type = "HttpNfcLease" [ 664.535170] env[65107]: } obtained for exporting VM: (result){ [ 664.535170] env[65107]: value = "vm-992635" [ 664.535170] env[65107]: _type = "VirtualMachine" [ 664.535170] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 664.535486] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the lease: (returnval){ [ 664.535486] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f9598-fbb8-29f4-8c58-4119fe5719f8" [ 664.535486] env[65107]: _type = "HttpNfcLease" [ 664.535486] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 664.551393] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102363, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.551393] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 664.551393] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f9598-fbb8-29f4-8c58-4119fe5719f8" [ 664.551393] env[65107]: _type = "HttpNfcLease" [ 664.551393] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 664.551664] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 664.551664] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f9598-fbb8-29f4-8c58-4119fe5719f8" [ 664.551664] env[65107]: _type = "HttpNfcLease" [ 664.551664] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 664.552176] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0fd1b0-1007-4a55-ae65-d7452e6a5399 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.568614] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529f91ea-aac9-0cc7-6bdc-323738df0ca5/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 664.568850] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529f91ea-aac9-0cc7-6bdc-323738df0ca5/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 664.579088] env[65107]: INFO nova.compute.manager [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Took 20.50 seconds to build instance. [ 664.640855] env[65107]: DEBUG oslo_concurrency.lockutils [None req-638c68c2-ab62-4539-a7d3-4154857eb5d9 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.584s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.642180] env[65107]: WARNING neutronclient.v2_0.client [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.642516] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.642665] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 664.642832] env[65107]: DEBUG nova.network.neutron [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 664.643013] env[65107]: DEBUG nova.objects.instance [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lazy-loading 'info_cache' on Instance uuid fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 664.679026] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-74dcb0d7-3dbd-4113-b11c-8404206849fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.754045] env[65107]: DEBUG oslo_concurrency.lockutils [None req-219ba123-c8fe-4704-aae9-10e89d42efd3 tempest-ServersAdmin275Test-189638262 tempest-ServersAdmin275Test-189638262-project-member] Lock "ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.864s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.758121] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.758509] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.947240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "7a3009bf-54a2-4565-a1aa-1d19286a4810" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.947240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.947240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "7a3009bf-54a2-4565-a1aa-1d19286a4810-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.947240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.947473] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.949306] env[65107]: INFO nova.compute.manager [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Terminating instance [ 664.991024] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52243b59-f11b-b690-783b-f23c0d9d2029, 'name': SearchDatastore_Task, 'duration_secs': 0.025341} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.991024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.991024] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 0e6c9f2d-4412-4d8d-8348-6552a34ab349/0e6c9f2d-4412-4d8d-8348-6552a34ab349.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 664.991024] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a134f4b0-af1f-4fe7-b3dd-e46d57cff084 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.004878] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 665.004878] env[65107]: value = "task-5102365" [ 665.004878] env[65107]: _type = "Task" [ 665.004878] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.018817] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.052681] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102363, 'name': Rename_Task, 'duration_secs': 0.352728} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.053303] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 665.053716] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9839aa8-4500-4a19-9bca-6edfa511a23b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.063436] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 665.063436] env[65107]: value = "task-5102366" [ 665.063436] env[65107]: _type = "Task" [ 665.063436] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.075699] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102366, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.155857] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6c4cde-676c-4762-865d-e241c164ba83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.167779] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d792d31d-6c05-4155-89f9-8eb5fe57ad4f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.206685] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee5d19e-07b9-4a40-96e3-13b5b2528b7e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.216119] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83699b18-672d-4aa0-b39b-a24d5d170aef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.233652] env[65107]: DEBUG nova.compute.provider_tree [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.271025] env[65107]: DEBUG nova.network.neutron [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Successfully updated port: 8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 665.464528] env[65107]: DEBUG nova.compute.manager [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 665.464962] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.465722] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a54632c-d723-4482-9b11-7f46c1973775 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.475869] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 665.477055] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29cd7405-3f39-487b-af95-8ea858ffb937 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.487325] env[65107]: DEBUG oslo_vmware.api [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 665.487325] env[65107]: value = "task-5102367" [ 665.487325] env[65107]: _type = "Task" [ 665.487325] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.501228] env[65107]: DEBUG oslo_vmware.api [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.520581] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102365, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.585651] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102366, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.592165] env[65107]: DEBUG nova.network.neutron [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 665.610244] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.610834] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.633019] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Received event network-vif-plugged-8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 665.633621] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Acquiring lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.633757] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.633839] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.633979] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] No waiting events found dispatching network-vif-plugged-8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 665.634162] env[65107]: WARNING nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Received unexpected event network-vif-plugged-8226470b-4e4b-4b3f-b40f-b29d3149af69 for instance with vm_state active and task_state None. [ 665.634346] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Received event network-changed-8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 665.634535] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Refreshing instance network info cache due to event network-changed-8226470b-4e4b-4b3f-b40f-b29d3149af69. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 665.634810] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Acquiring lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.635027] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Acquired lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.635543] env[65107]: DEBUG nova.network.neutron [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Refreshing network info cache for port 8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 665.640140] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.640833] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.662322] env[65107]: WARNING neutronclient.v2_0.client [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.662711] env[65107]: WARNING openstack [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.663034] env[65107]: WARNING openstack [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.739131] env[65107]: DEBUG nova.scheduler.client.report [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 665.774573] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "refresh_cache-f7446d76-45c2-4e8b-981d-d37c230cf125" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.775119] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquired lock "refresh_cache-f7446d76-45c2-4e8b-981d-d37c230cf125" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.775308] env[65107]: DEBUG nova.network.neutron [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 665.784786] env[65107]: WARNING neutronclient.v2_0.client [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.785798] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.786292] env[65107]: WARNING openstack [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.858154] env[65107]: WARNING neutronclient.v2_0.client [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.859244] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.859722] env[65107]: WARNING openstack [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.001419] env[65107]: DEBUG oslo_vmware.api [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102367, 'name': PowerOffVM_Task, 'duration_secs': 0.369494} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.008192] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 666.008192] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 666.008724] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e7e018d-86a4-49bc-9c9e-52bcbd299c14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.022041] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690248} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.022342] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 0e6c9f2d-4412-4d8d-8348-6552a34ab349/0e6c9f2d-4412-4d8d-8348-6552a34ab349.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 666.022573] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 666.022937] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0be54f6d-5ae0-4d76-a19e-8ad6b0ce1573 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.030522] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 666.030522] env[65107]: value = "task-5102369" [ 666.030522] env[65107]: _type = "Task" [ 666.030522] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.046835] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102369, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.080102] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102366, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.103204] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 666.103492] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 666.103716] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Deleting the datastore file [datastore1] 7a3009bf-54a2-4565-a1aa-1d19286a4810 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 666.104044] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e49666ad-ad51-4c0d-add0-55ec36ffb173 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.113401] env[65107]: DEBUG oslo_vmware.api [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for the task: (returnval){ [ 666.113401] env[65107]: value = "task-5102370" [ 666.113401] env[65107]: _type = "Task" [ 666.113401] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.124601] env[65107]: DEBUG oslo_vmware.api [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102370, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.140960] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Updated VIF entry in instance network info cache for port 2275d82f-e805-4c7b-8656-ff9475e7f524. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 666.141526] env[65107]: DEBUG nova.network.neutron [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Updating instance_info_cache with network_info: [{"id": "2275d82f-e805-4c7b-8656-ff9475e7f524", "address": "fa:16:3e:67:1e:59", "network": {"id": "c165d490-0569-4946-966c-3a5820469066", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1542494870-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "520207ef29e64aa799841fae36e469f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2275d82f-e8", "ovs_interfaceid": "2275d82f-e805-4c7b-8656-ff9475e7f524", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.143797] env[65107]: WARNING neutronclient.v2_0.client [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.144561] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.145331] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.198076] env[65107]: WARNING openstack [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.198730] env[65107]: WARNING openstack [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.244739] env[65107]: DEBUG nova.network.neutron [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updating instance_info_cache with network_info: [{"id": "09df7af1-1609-4648-a7d1-a870687d5bbb", "address": "fa:16:3e:d9:57:dd", "network": {"id": "de1c1b5c-0632-45e3-9134-5fbc553a0cf0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-842310743-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3fa2b26480f4495926fadf02489de3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09df7af1-16", "ovs_interfaceid": "09df7af1-1609-4648-a7d1-a870687d5bbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.247468] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.030s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.251495] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.193s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.251779] env[65107]: DEBUG nova.objects.instance [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lazy-loading 'resources' on Instance uuid 4e70aaf3-5f1c-4a61-a790-a0a5e409170d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 666.278364] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.279920] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.294718] env[65107]: INFO nova.scheduler.client.report [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted allocations for instance abb9dc0e-ce18-49c8-a472-d39a3fcec887 [ 666.454518] env[65107]: DEBUG nova.network.neutron [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 666.540916] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102369, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080511} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.545180] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 666.545180] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8117fb-9375-4a11-be2f-e847668c128f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.579608] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] 0e6c9f2d-4412-4d8d-8348-6552a34ab349/0e6c9f2d-4412-4d8d-8348-6552a34ab349.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 666.581224] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.581224] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.590535] env[65107]: WARNING neutronclient.v2_0.client [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.591134] env[65107]: WARNING openstack [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.591649] env[65107]: WARNING openstack [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.604350] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97bc2da0-23a7-487a-98b7-c1b4a747b641 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.622560] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.623210] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.644819] env[65107]: DEBUG oslo_vmware.api [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102366, 'name': PowerOnVM_Task, 'duration_secs': 1.149981} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.649922] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 666.650251] env[65107]: INFO nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Took 13.02 seconds to spawn the instance on the hypervisor. [ 666.651860] env[65107]: DEBUG nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 666.651860] env[65107]: DEBUG oslo_vmware.api [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Task: {'id': task-5102370, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361564} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.652257] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 666.652257] env[65107]: value = "task-5102371" [ 666.652257] env[65107]: _type = "Task" [ 666.652257] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.653940] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fb574c-97b8-4426-9dc9-206bf2cd7180 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.656558] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 666.657398] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 666.657398] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.658294] env[65107]: INFO nova.compute.manager [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Took 1.19 seconds to destroy the instance on the hypervisor. [ 666.658603] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 666.659209] env[65107]: DEBUG oslo_concurrency.lockutils [req-7f0b012c-9922-4123-a4de-8d7fc4899adc req-27e54167-6bf7-48fd-a23c-0810ffd1348d service nova] Releasing lock "refresh_cache-e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.659757] env[65107]: DEBUG nova.compute.manager [-] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 666.661713] env[65107]: DEBUG nova.network.neutron [-] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 666.661713] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.662900] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.663962] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.709312] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102371, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.752027] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Releasing lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.752027] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Instance network_info: |[{"id": "09df7af1-1609-4648-a7d1-a870687d5bbb", "address": "fa:16:3e:d9:57:dd", "network": {"id": "de1c1b5c-0632-45e3-9134-5fbc553a0cf0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-842310743-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3fa2b26480f4495926fadf02489de3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09df7af1-16", "ovs_interfaceid": "09df7af1-1609-4648-a7d1-a870687d5bbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 666.752573] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:57:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '09df7af1-1609-4648-a7d1-a870687d5bbb', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.764434] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Creating folder: Project (e3fa2b26480f4495926fadf02489de3e). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.766996] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2fdb49a7-2ca7-4118-8c1e-24a7b8c31288 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.783946] env[65107]: WARNING neutronclient.v2_0.client [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.785197] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.786496] env[65107]: WARNING openstack [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.799220] env[65107]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 666.800062] env[65107]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65107) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 666.800682] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Folder already exists: Project (e3fa2b26480f4495926fadf02489de3e). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 666.801502] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Creating folder: Instances. Parent ref: group-v992605. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.804955] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b000156f-a807-4140-bf0a-3f7a35702959 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.812550] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.816836] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b48fad38-d371-436e-99ab-81cc59ee0f21 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "abb9dc0e-ce18-49c8-a472-d39a3fcec887" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.207s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.824599] env[65107]: DEBUG nova.network.neutron [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance_info_cache with network_info: [{"id": "086361c0-3cd7-4389-b5f0-c550a2adc834", "address": "fa:16:3e:bf:a0:a1", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086361c0-3c", "ovs_interfaceid": "086361c0-3cd7-4389-b5f0-c550a2adc834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.831991] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Created folder: Instances in parent group-v992605. [ 666.832518] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 666.832646] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7062b119-01a4-45f4-bb52-212178382271] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 666.832974] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20c6e8e6-cc8d-42a5-a727-acce395c3055 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.852068] env[65107]: WARNING neutronclient.v2_0.client [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.852931] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.853452] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.873901] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.873901] env[65107]: value = "task-5102374" [ 666.873901] env[65107]: _type = "Task" [ 666.873901] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.890117] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102374, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.077085] env[65107]: DEBUG nova.compute.manager [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Received event network-vif-plugged-55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 667.077498] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Acquiring lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.077642] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.082022] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.082022] env[65107]: DEBUG nova.compute.manager [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] No waiting events found dispatching network-vif-plugged-55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 667.082604] env[65107]: WARNING nova.compute.manager [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Received unexpected event network-vif-plugged-55102692-2502-4633-b00b-cc29760fb310 for instance with vm_state building and task_state spawning. [ 667.082604] env[65107]: DEBUG nova.compute.manager [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Received event network-changed-55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 667.082769] env[65107]: DEBUG nova.compute.manager [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Refreshing instance network info cache due to event network-changed-55102692-2502-4633-b00b-cc29760fb310. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 667.082888] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Acquiring lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.082956] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Acquired lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.083200] env[65107]: DEBUG nova.network.neutron [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Refreshing network info cache for port 55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 667.114648] env[65107]: DEBUG nova.network.neutron [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Updating instance_info_cache with network_info: [{"id": "8a1e7b3c-e981-43b6-84a0-1adcd2b2a044", "address": "fa:16:3e:a1:79:24", "network": {"id": "0490082a-0acf-460a-9ea8-c4b5bc26ed08", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1324351848-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95393e0b56bd47cd8bb37207e0500269", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a1e7b3c-e9", "ovs_interfaceid": "8a1e7b3c-e981-43b6-84a0-1adcd2b2a044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.167047] env[65107]: DEBUG nova.network.neutron [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Updated VIF entry in instance network info cache for port 8226470b-4e4b-4b3f-b40f-b29d3149af69. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 667.167190] env[65107]: DEBUG nova.network.neutron [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Updating instance_info_cache with network_info: [{"id": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "address": "fa:16:3e:92:8d:9e", "network": {"id": "b9e7c478-0189-4385-98a5-bc3ca80526e5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1623056907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c5372cbba21404eb54eda5f08f0ac42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea6e81c3-94aa-40a6-a4d4-7f338b503442", "external-id": "nsx-vlan-transportzone-637", "segmentation_id": 637, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8226470b-4e", "ovs_interfaceid": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.179369] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102371, 'name': ReconfigVM_Task, 'duration_secs': 0.391655} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.179369] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Reconfigured VM instance instance-00000016 to attach disk [datastore1] 0e6c9f2d-4412-4d8d-8348-6552a34ab349/0e6c9f2d-4412-4d8d-8348-6552a34ab349.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 667.179369] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd2c979c-06ee-481b-a02f-bf04536cd598 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.191308] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 667.191308] env[65107]: value = "task-5102375" [ 667.191308] env[65107]: _type = "Task" [ 667.191308] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.191894] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.196108] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.211250] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102375, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.219538] env[65107]: INFO nova.compute.manager [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Took 22.34 seconds to build instance. [ 667.310910] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43ccf9f-4e70-4a31-a2e6-7a55731f730e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.324544] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25baf36f-4a9e-4d5d-9c00-26aa700927e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.331406] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.331406] env[65107]: DEBUG nova.objects.instance [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lazy-loading 'migration_context' on Instance uuid fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 667.366464] env[65107]: DEBUG nova.objects.base [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 667.367445] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce2ab59-c5eb-42c0-bb9a-ee8799fcf44b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.371547] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb3ff58-11bd-4fda-8c7e-7e792d7a00c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.400854] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7620090-5018-4f8c-b7a7-d1586a82e981 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.409697] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc02cebc-7457-46c9-8525-becf422231c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.412216] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102374, 'name': CreateVM_Task, 'duration_secs': 0.455691} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.412837] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7062b119-01a4-45f4-bb52-212178382271] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 667.413722] env[65107]: WARNING neutronclient.v2_0.client [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.414162] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992609', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'name': 'volume-7ab345f6-458d-4bb0-be22-c593a481ce3f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7062b119-01a4-45f4-bb52-212178382271', 'attached_at': '', 'detached_at': '', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'serial': '7ab345f6-458d-4bb0-be22-c593a481ce3f'}, 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'attachment_id': 'b97e3c6b-a27d-4c54-a315-be3047baf127', 'mount_device': '/dev/sda', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=65107) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 667.414323] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Root volume attach. Driver type: vmdk {{(pid=65107) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 667.415174] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422d7eb2-6016-405f-a3e3-42f1f2cddec3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.431043] env[65107]: DEBUG oslo_vmware.api [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 667.431043] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521a5bcb-3d95-ef31-e5a8-3903b12523e7" [ 667.431043] env[65107]: _type = "Task" [ 667.431043] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.431043] env[65107]: DEBUG nova.compute.provider_tree [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.441602] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795a7eae-ac19-42e6-a63a-f2e07274741e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.452338] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "b20a2c71-514d-488a-8edb-8fb0274a894a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.452338] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.454634] env[65107]: DEBUG oslo_vmware.api [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521a5bcb-3d95-ef31-e5a8-3903b12523e7, 'name': SearchDatastore_Task, 'duration_secs': 0.008793} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.454634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.458589] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b74bb67-a7ef-4b3d-bb10-7593e440d819 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.466205] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-a914088a-9f86-4c00-ab6d-4a48a037602c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.476067] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 667.476067] env[65107]: value = "task-5102376" [ 667.476067] env[65107]: _type = "Task" [ 667.476067] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.486939] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102376, 'name': RelocateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.590661] env[65107]: WARNING neutronclient.v2_0.client [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.591673] env[65107]: WARNING openstack [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.591787] env[65107]: WARNING openstack [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.618381] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Releasing lock "refresh_cache-f7446d76-45c2-4e8b-981d-d37c230cf125" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.618381] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Instance network_info: |[{"id": "8a1e7b3c-e981-43b6-84a0-1adcd2b2a044", "address": "fa:16:3e:a1:79:24", "network": {"id": "0490082a-0acf-460a-9ea8-c4b5bc26ed08", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1324351848-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95393e0b56bd47cd8bb37207e0500269", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a1e7b3c-e9", "ovs_interfaceid": "8a1e7b3c-e981-43b6-84a0-1adcd2b2a044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 667.618692] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:79:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca401eaa-889a-4f9f-ac9a-56b4c41bfc06', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8a1e7b3c-e981-43b6-84a0-1adcd2b2a044', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 667.627014] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Creating folder: Project (95393e0b56bd47cd8bb37207e0500269). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.628709] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bc83f60-749e-4d2f-bc89-9bc7ea48bb3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.634595] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.634838] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.647530] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Created folder: Project (95393e0b56bd47cd8bb37207e0500269) in parent group-v992574. [ 667.647663] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Creating folder: Instances. Parent ref: group-v992641. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.648791] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-050e166a-ff90-4377-8a95-f3f69a966ff0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.662729] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Created folder: Instances in parent group-v992641. [ 667.662729] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 667.662729] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 667.662908] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3619517-46d8-456a-8a13-9027d5933d1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.679772] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Releasing lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.680073] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Received event network-vif-plugged-d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 667.680352] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Acquiring lock "7b24c530-56bf-4666-96f9-e2dc66728f1a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.680464] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.680648] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.680833] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] No waiting events found dispatching network-vif-plugged-d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 667.681045] env[65107]: WARNING nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Received unexpected event network-vif-plugged-d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff for instance with vm_state building and task_state spawning. [ 667.681319] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Received event network-changed-d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 667.681602] env[65107]: DEBUG nova.compute.manager [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Refreshing instance network info cache due to event network-changed-d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 667.681918] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Acquiring lock "refresh_cache-7b24c530-56bf-4666-96f9-e2dc66728f1a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.682162] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Acquired lock "refresh_cache-7b24c530-56bf-4666-96f9-e2dc66728f1a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.682446] env[65107]: DEBUG nova.network.neutron [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Refreshing network info cache for port d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 667.700541] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 667.700541] env[65107]: value = "task-5102379" [ 667.700541] env[65107]: _type = "Task" [ 667.700541] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.727316] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.731474] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7a9dff8d-92be-49df-bf5b-9f80c83c55d0 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.864s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.731979] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.732396] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102375, 'name': Rename_Task, 'duration_secs': 0.209215} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.734717] env[65107]: WARNING openstack [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.735457] env[65107]: WARNING openstack [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.749600] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.749991] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 667.750997] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.751234] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ae651fe-1aa2-4582-8ab1-b45ac922d583 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.758800] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.759572] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102379, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.759793] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.759958] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 667.760841] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 667.766783] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 667.766783] env[65107]: value = "task-5102380" [ 667.766783] env[65107]: _type = "Task" [ 667.766783] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.772346] env[65107]: DEBUG nova.network.neutron [-] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.782082] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.815082] env[65107]: WARNING neutronclient.v2_0.client [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.815323] env[65107]: WARNING openstack [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.815560] env[65107]: WARNING openstack [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.935293] env[65107]: DEBUG nova.scheduler.client.report [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 667.957093] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 667.994765] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102376, 'name': RelocateVM_Task} progress is 20%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.003296] env[65107]: DEBUG nova.network.neutron [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Updated VIF entry in instance network info cache for port 55102692-2502-4633-b00b-cc29760fb310. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 668.003760] env[65107]: DEBUG nova.network.neutron [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Updating instance_info_cache with network_info: [{"id": "55102692-2502-4633-b00b-cc29760fb310", "address": "fa:16:3e:6a:99:59", "network": {"id": "90ceb5af-809c-4b48-9c35-26e48235753c", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-156477542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e9fcd71f4f064b488f832406fedafbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55102692-25", "ovs_interfaceid": "55102692-2502-4633-b00b-cc29760fb310", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 668.137583] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 668.186452] env[65107]: WARNING neutronclient.v2_0.client [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.188288] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.188890] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.223761] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102379, 'name': CreateVM_Task, 'duration_secs': 0.426939} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.224268] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 668.225343] env[65107]: WARNING neutronclient.v2_0.client [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.225741] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.225871] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.226257] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 668.226615] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06371e17-08a5-41f2-9efd-4fd67f044138 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.233921] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 668.233921] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526986b3-8041-5846-7ea5-b225fbe51c93" [ 668.233921] env[65107]: _type = "Task" [ 668.233921] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.245991] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526986b3-8041-5846-7ea5-b225fbe51c93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.264304] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.278764] env[65107]: INFO nova.compute.manager [-] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Took 1.62 seconds to deallocate network for instance. [ 668.279219] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102380, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.447282] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.448309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.932s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.449101] env[65107]: DEBUG nova.objects.instance [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'resources' on Instance uuid 0691bdfe-ff23-471b-ad94-2a98adcecd4f {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 668.489749] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.491661] env[65107]: INFO nova.scheduler.client.report [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Deleted allocations for instance 4e70aaf3-5f1c-4a61-a790-a0a5e409170d [ 668.505073] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102376, 'name': RelocateVM_Task, 'duration_secs': 0.71921} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.505981] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 668.509058] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992609', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'name': 'volume-7ab345f6-458d-4bb0-be22-c593a481ce3f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7062b119-01a4-45f4-bb52-212178382271', 'attached_at': '', 'detached_at': '', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'serial': '7ab345f6-458d-4bb0-be22-c593a481ce3f'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 668.518070] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0e2491f-4a5b-439b-b781-86ebbfb77bc8 req-be759ed1-87e3-44cd-9ba9-70250b68b06f service nova] Releasing lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.519840] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c045adeb-41c2-410d-abab-2abf05d79a5a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.551312] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75b3518-0358-4025-a805-2f5fbc55dd25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.591324] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] volume-7ab345f6-458d-4bb0-be22-c593a481ce3f/volume-7ab345f6-458d-4bb0-be22-c593a481ce3f.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 668.592834] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b669c327-e010-43fd-84ec-9d19d74812b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.627495] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 668.627495] env[65107]: value = "task-5102381" [ 668.627495] env[65107]: _type = "Task" [ 668.627495] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.649079] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102381, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.668632] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.745084] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526986b3-8041-5846-7ea5-b225fbe51c93, 'name': SearchDatastore_Task, 'duration_secs': 0.020017} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.745460] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.745666] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 668.745920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.746068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.746264] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 668.746548] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6763e837-ca2b-4435-995d-cdc9e3c2e9f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.755136] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.755542] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.764210] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 668.764401] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 668.765136] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddea6228-27d3-495b-8382-de096de5d5a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.773546] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 668.773546] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2b613-046b-6dc1-1740-a0a8cf6e43be" [ 668.773546] env[65107]: _type = "Task" [ 668.773546] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.781477] env[65107]: DEBUG oslo_vmware.api [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102380, 'name': PowerOnVM_Task, 'duration_secs': 0.740351} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.782215] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 668.782432] env[65107]: INFO nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Took 12.28 seconds to spawn the instance on the hypervisor. [ 668.782610] env[65107]: DEBUG nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 668.783435] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3e0b1d-e997-4c0d-a2f0-eb834284de09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.790380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.790647] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2b613-046b-6dc1-1740-a0a8cf6e43be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.009536] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f08d38d2-a358-48fd-a399-6234643be16f tempest-InstanceActionsTestJSON-1305569336 tempest-InstanceActionsTestJSON-1305569336-project-member] Lock "4e70aaf3-5f1c-4a61-a790-a0a5e409170d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.988s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.134367] env[65107]: WARNING neutronclient.v2_0.client [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.135727] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.135898] env[65107]: WARNING openstack [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.155979] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102381, 'name': ReconfigVM_Task, 'duration_secs': 0.474668} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.155979] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Reconfigured VM instance instance-00000017 to attach disk [datastore2] volume-7ab345f6-458d-4bb0-be22-c593a481ce3f/volume-7ab345f6-458d-4bb0-be22-c593a481ce3f.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 669.164945] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fcb8267-eca3-4315-8978-4d552f8f4543 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.183290] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 669.183290] env[65107]: value = "task-5102382" [ 669.183290] env[65107]: _type = "Task" [ 669.183290] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.198045] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102382, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.286049] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2b613-046b-6dc1-1740-a0a8cf6e43be, 'name': SearchDatastore_Task, 'duration_secs': 0.016824} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.290197] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb7377df-2335-4600-9419-44414fd6bdd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.300664] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 669.300664] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5246046b-99c2-b8bf-1e0e-acafdfb55fda" [ 669.300664] env[65107]: _type = "Task" [ 669.300664] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.322042] env[65107]: INFO nova.compute.manager [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Took 23.31 seconds to build instance. [ 669.322533] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5246046b-99c2-b8bf-1e0e-acafdfb55fda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.404948] env[65107]: DEBUG nova.compute.manager [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Received event network-vif-plugged-8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 669.407237] env[65107]: DEBUG oslo_concurrency.lockutils [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Acquiring lock "f7446d76-45c2-4e8b-981d-d37c230cf125-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.407530] env[65107]: DEBUG oslo_concurrency.lockutils [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.408437] env[65107]: DEBUG oslo_concurrency.lockutils [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.408590] env[65107]: DEBUG nova.compute.manager [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] No waiting events found dispatching network-vif-plugged-8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 669.408760] env[65107]: WARNING nova.compute.manager [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Received unexpected event network-vif-plugged-8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 for instance with vm_state building and task_state spawning. [ 669.410256] env[65107]: DEBUG nova.compute.manager [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Received event network-changed-8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 669.410462] env[65107]: DEBUG nova.compute.manager [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Refreshing instance network info cache due to event network-changed-8a1e7b3c-e981-43b6-84a0-1adcd2b2a044. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 669.410666] env[65107]: DEBUG oslo_concurrency.lockutils [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Acquiring lock "refresh_cache-f7446d76-45c2-4e8b-981d-d37c230cf125" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.410803] env[65107]: DEBUG oslo_concurrency.lockutils [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Acquired lock "refresh_cache-f7446d76-45c2-4e8b-981d-d37c230cf125" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.410957] env[65107]: DEBUG nova.network.neutron [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Refreshing network info cache for port 8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 669.438948] env[65107]: DEBUG nova.network.neutron [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Updated VIF entry in instance network info cache for port d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 669.439711] env[65107]: DEBUG nova.network.neutron [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Updating instance_info_cache with network_info: [{"id": "d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff", "address": "fa:16:3e:43:32:e9", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.27", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd533dd9c-4b", "ovs_interfaceid": "d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.442798] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bfb13c-0874-4b0f-aa47-f5037520e9c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.453156] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a4c6a8-5169-41b9-ac47-9c1cad2501fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.496836] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477a4be4-6ce2-465a-8160-d507b7c07cf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.508815] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f10258d-35e2-4028-af32-cf347a7d72df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.526164] env[65107]: DEBUG nova.compute.provider_tree [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.698924] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102382, 'name': ReconfigVM_Task, 'duration_secs': 0.179116} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.700475] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992609', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'name': 'volume-7ab345f6-458d-4bb0-be22-c593a481ce3f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7062b119-01a4-45f4-bb52-212178382271', 'attached_at': '', 'detached_at': '', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'serial': '7ab345f6-458d-4bb0-be22-c593a481ce3f'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 669.701332] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96c62e48-3281-4b6e-9792-6bb2ee59f3b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.710750] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 669.710750] env[65107]: value = "task-5102383" [ 669.710750] env[65107]: _type = "Task" [ 669.710750] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.721517] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102383, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.811149] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5246046b-99c2-b8bf-1e0e-acafdfb55fda, 'name': SearchDatastore_Task, 'duration_secs': 0.024688} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.811693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.812148] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] f7446d76-45c2-4e8b-981d-d37c230cf125/f7446d76-45c2-4e8b-981d-d37c230cf125.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 669.812492] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e94c0202-c578-49ee-a648-1c3bda454093 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.823315] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 669.823315] env[65107]: value = "task-5102384" [ 669.823315] env[65107]: _type = "Task" [ 669.823315] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.825033] env[65107]: DEBUG oslo_concurrency.lockutils [None req-645e231b-ec58-454b-be84-4c4a1681a2d3 tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.821s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.837025] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.921065] env[65107]: WARNING neutronclient.v2_0.client [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.922122] env[65107]: WARNING openstack [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.922599] env[65107]: WARNING openstack [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.947642] env[65107]: DEBUG oslo_concurrency.lockutils [req-51be0983-c275-4d41-864e-49f1a4a3c37e req-5c7b56e3-9fa0-44ea-9104-e166a3198de9 service nova] Releasing lock "refresh_cache-7b24c530-56bf-4666-96f9-e2dc66728f1a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.032448] env[65107]: DEBUG nova.scheduler.client.report [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 670.226216] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102383, 'name': Rename_Task, 'duration_secs': 0.210202} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.226450] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 670.227048] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b91243d-c174-4572-9628-9441638fd1a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.237259] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 670.237259] env[65107]: value = "task-5102386" [ 670.237259] env[65107]: _type = "Task" [ 670.237259] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.250955] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.339311] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102384, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.417740] env[65107]: WARNING openstack [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.418377] env[65107]: WARNING openstack [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.543033] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.090s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.545368] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.549s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.548570] env[65107]: INFO nova.compute.claims [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.594356] env[65107]: INFO nova.scheduler.client.report [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocations for instance 0691bdfe-ff23-471b-ad94-2a98adcecd4f [ 670.628123] env[65107]: WARNING neutronclient.v2_0.client [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.628123] env[65107]: WARNING openstack [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.628123] env[65107]: WARNING openstack [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.750929] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102386, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.835993] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102384, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64514} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.842294] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] f7446d76-45c2-4e8b-981d-d37c230cf125/f7446d76-45c2-4e8b-981d-d37c230cf125.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 670.842695] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 670.843436] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d51080a-6181-4038-a5fa-84bfcfbca7cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.854467] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 670.854467] env[65107]: value = "task-5102389" [ 670.854467] env[65107]: _type = "Task" [ 670.854467] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.868857] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.105570] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d1fbc1a-6f36-47d0-8d5a-8232bc1cb370 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "0691bdfe-ff23-471b-ad94-2a98adcecd4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.144s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.210535] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "07b83b81-9c65-4180-b208-c96525b78d8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.210535] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "07b83b81-9c65-4180-b208-c96525b78d8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.246222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.246486] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.254837] env[65107]: DEBUG oslo_vmware.api [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102386, 'name': PowerOnVM_Task, 'duration_secs': 0.793724} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.255303] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 671.255641] env[65107]: INFO nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Took 7.77 seconds to spawn the instance on the hypervisor. [ 671.255968] env[65107]: DEBUG nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 671.256891] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ef15a4-9b64-4cb3-be63-6ccc969141d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.365865] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.657088] env[65107]: DEBUG nova.network.neutron [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Updated VIF entry in instance network info cache for port 8a1e7b3c-e981-43b6-84a0-1adcd2b2a044. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 671.658164] env[65107]: DEBUG nova.network.neutron [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Updating instance_info_cache with network_info: [{"id": "8a1e7b3c-e981-43b6-84a0-1adcd2b2a044", "address": "fa:16:3e:a1:79:24", "network": {"id": "0490082a-0acf-460a-9ea8-c4b5bc26ed08", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1324351848-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95393e0b56bd47cd8bb37207e0500269", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca401eaa-889a-4f9f-ac9a-56b4c41bfc06", "external-id": "nsx-vlan-transportzone-877", "segmentation_id": 877, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a1e7b3c-e9", "ovs_interfaceid": "8a1e7b3c-e981-43b6-84a0-1adcd2b2a044", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 671.720218] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 671.727907] env[65107]: DEBUG nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Received event network-vif-plugged-09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 671.728405] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Acquiring lock "7062b119-01a4-45f4-bb52-212178382271-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.729033] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Lock "7062b119-01a4-45f4-bb52-212178382271-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.729033] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Lock "7062b119-01a4-45f4-bb52-212178382271-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.730088] env[65107]: DEBUG nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] No waiting events found dispatching network-vif-plugged-09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 671.730351] env[65107]: WARNING nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Received unexpected event network-vif-plugged-09df7af1-1609-4648-a7d1-a870687d5bbb for instance with vm_state active and task_state None. [ 671.730757] env[65107]: DEBUG nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Received event network-changed-09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 671.730919] env[65107]: DEBUG nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Refreshing instance network info cache due to event network-changed-09df7af1-1609-4648-a7d1-a870687d5bbb. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 671.731323] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Acquiring lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.731739] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Acquired lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.733462] env[65107]: DEBUG nova.network.neutron [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Refreshing network info cache for port 09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 671.749708] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 671.781378] env[65107]: INFO nova.compute.manager [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Took 22.01 seconds to build instance. [ 671.874884] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.854362} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.875311] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 671.877426] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a79de3-936d-4edc-9fe0-b5f48be296d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.905323] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] f7446d76-45c2-4e8b-981d-d37c230cf125/f7446d76-45c2-4e8b-981d-d37c230cf125.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 671.906770] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b897e309-fbae-42f2-8c50-6386b6a20567 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.934579] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 671.934579] env[65107]: value = "task-5102390" [ 671.934579] env[65107]: _type = "Task" [ 671.934579] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.950050] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.142246] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe30120-cc83-47be-9a49-b0ecbccf97c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.151924] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845264d0-b94b-410f-805f-08b58d07a049 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.191834] env[65107]: DEBUG oslo_concurrency.lockutils [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] Releasing lock "refresh_cache-f7446d76-45c2-4e8b-981d-d37c230cf125" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.192221] env[65107]: DEBUG nova.compute.manager [req-76fe4592-567f-48f0-94f4-acfa7ad1ddff req-1cf93f74-55e3-4f79-b6db-fe192a928b88 service nova] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Received event network-vif-deleted-20915f30-cfbb-4a09-bf64-495545e0096f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 672.194028] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749b59ce-44eb-45e9-bb9f-c41978b4f35e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.203788] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428a8547-19db-4a1e-8a1b-d828955b75a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.228781] env[65107]: DEBUG nova.compute.provider_tree [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.238433] env[65107]: WARNING neutronclient.v2_0.client [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.239162] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.240200] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.274181] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.285838] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.286778] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0d5a71d4-dc76-4811-bc86-45e3328ce759 tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "7062b119-01a4-45f4-bb52-212178382271" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.532s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.446916] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.733899] env[65107]: DEBUG nova.scheduler.client.report [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 672.853350] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.854150] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.952915] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.135150] env[65107]: WARNING neutronclient.v2_0.client [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.135150] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.135686] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.241275] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.696s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.242049] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 673.246249] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.808s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 673.248802] env[65107]: INFO nova.compute.claims [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.450647] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102390, 'name': ReconfigVM_Task, 'duration_secs': 1.265455} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.451481] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Reconfigured VM instance instance-00000018 to attach disk [datastore2] f7446d76-45c2-4e8b-981d-d37c230cf125/f7446d76-45c2-4e8b-981d-d37c230cf125.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 673.451776] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8b54a84-d012-447d-8ebb-83c82cb28eb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.462599] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 673.462599] env[65107]: value = "task-5102392" [ 673.462599] env[65107]: _type = "Task" [ 673.462599] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.479036] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102392, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.508493] env[65107]: DEBUG nova.network.neutron [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updated VIF entry in instance network info cache for port 09df7af1-1609-4648-a7d1-a870687d5bbb. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 673.509148] env[65107]: DEBUG nova.network.neutron [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updating instance_info_cache with network_info: [{"id": "09df7af1-1609-4648-a7d1-a870687d5bbb", "address": "fa:16:3e:d9:57:dd", "network": {"id": "de1c1b5c-0632-45e3-9134-5fbc553a0cf0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-842310743-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3fa2b26480f4495926fadf02489de3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09df7af1-16", "ovs_interfaceid": "09df7af1-1609-4648-a7d1-a870687d5bbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.545431] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529f91ea-aac9-0cc7-6bdc-323738df0ca5/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 673.545431] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf5e591-75b3-4493-b57f-718f9f3db1e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.552342] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529f91ea-aac9-0cc7-6bdc-323738df0ca5/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 673.552543] env[65107]: ERROR oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529f91ea-aac9-0cc7-6bdc-323738df0ca5/disk-0.vmdk due to incomplete transfer. [ 673.552791] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-48dc4059-150a-40e3-9f88-1618c0f3e6fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.563287] env[65107]: DEBUG oslo_vmware.rw_handles [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529f91ea-aac9-0cc7-6bdc-323738df0ca5/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 673.563580] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Uploaded image a656a63b-c667-4970-863b-5666f3f5df1d to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 673.566099] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 673.566788] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c12f7a99-0106-45fa-a011-a27ab9549a2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.574211] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 673.574211] env[65107]: value = "task-5102393" [ 673.574211] env[65107]: _type = "Task" [ 673.574211] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.583878] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102393, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.756058] env[65107]: DEBUG nova.compute.utils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 673.765166] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 673.765166] env[65107]: DEBUG nova.network.neutron [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 673.765166] env[65107]: WARNING neutronclient.v2_0.client [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.765166] env[65107]: WARNING neutronclient.v2_0.client [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.765166] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.765446] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.947546] env[65107]: DEBUG nova.policy [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd7f9abea8d946ecb2accee2714703e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2d4844feda0432ebe1ef160b0070b83', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 673.976839] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102392, 'name': Rename_Task, 'duration_secs': 0.323858} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.977163] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 673.977497] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-973c2f2c-4875-4946-b88b-b0cc95a27143 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.988155] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 673.988155] env[65107]: value = "task-5102394" [ 673.988155] env[65107]: _type = "Task" [ 673.988155] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.000174] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.018801] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Releasing lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.018972] env[65107]: DEBUG nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 674.019296] env[65107]: DEBUG nova.compute.manager [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing instance network info cache due to event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 674.021576] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.021576] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.021576] env[65107]: DEBUG nova.network.neutron [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 674.087628] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102393, 'name': Destroy_Task, 'duration_secs': 0.444473} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.087857] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Destroyed the VM [ 674.088281] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 674.088498] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ff446d47-2595-418a-a6a5-a4c61f9d68ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.098334] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 674.098334] env[65107]: value = "task-5102395" [ 674.098334] env[65107]: _type = "Task" [ 674.098334] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.108966] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102395, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.277376] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 674.335209] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755c925a-8239-4c2c-b380-df700c506557 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.346653] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c128e7-82a8-4a4b-9042-463d65b84cb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.388627] env[65107]: DEBUG nova.network.neutron [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Successfully created port: 8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 674.391647] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfa071a-50e2-4f60-8bcf-1908ad4f7e6d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.401094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8234ed4-d386-4475-82ee-49d1fd9d99be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.419172] env[65107]: DEBUG nova.compute.provider_tree [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.502198] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102394, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.524785] env[65107]: WARNING neutronclient.v2_0.client [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.526228] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.526504] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.610629] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102395, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.733945] env[65107]: DEBUG nova.compute.manager [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Received event network-changed-02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 674.734166] env[65107]: DEBUG nova.compute.manager [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Refreshing instance network info cache due to event network-changed-02108638-8e1e-447a-9b94-a45adfd32161. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 674.734366] env[65107]: DEBUG oslo_concurrency.lockutils [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Acquiring lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.735909] env[65107]: DEBUG oslo_concurrency.lockutils [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Acquired lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.735909] env[65107]: DEBUG nova.network.neutron [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Refreshing network info cache for port 02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 674.924411] env[65107]: DEBUG nova.scheduler.client.report [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 675.000533] env[65107]: DEBUG oslo_vmware.api [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102394, 'name': PowerOnVM_Task, 'duration_secs': 0.960883} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.000793] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 675.000987] env[65107]: INFO nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Took 10.82 seconds to spawn the instance on the hypervisor. [ 675.001247] env[65107]: DEBUG nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 675.003691] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad1eabb-2c3b-43dd-aef2-3a7c4d2604c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.111569] env[65107]: DEBUG oslo_vmware.api [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102395, 'name': RemoveSnapshot_Task, 'duration_secs': 0.622645} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.111569] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 675.111569] env[65107]: INFO nova.compute.manager [None req-79250d22-5deb-4c2a-8b41-029a3594ce53 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Took 14.77 seconds to snapshot the instance on the hypervisor. [ 675.242017] env[65107]: WARNING neutronclient.v2_0.client [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.242759] env[65107]: WARNING openstack [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.243109] env[65107]: WARNING openstack [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.289414] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 675.321407] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 675.321669] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 675.321825] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 675.322013] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 675.322339] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 675.322625] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 675.322797] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.322988] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 675.323217] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 675.323446] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 675.323664] env[65107]: DEBUG nova.virt.hardware [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 675.324689] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3002bb3-ea0e-434d-9a8c-f190555fcefc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.333830] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014a4062-d805-4f75-8797-ef203587ad90 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.430246] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.184s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.430377] env[65107]: DEBUG nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 675.434516] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.979s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.526862] env[65107]: INFO nova.compute.manager [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Took 25.75 seconds to build instance. [ 675.938671] env[65107]: DEBUG nova.compute.utils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 675.940700] env[65107]: DEBUG nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 676.030116] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34a9f6a4-7734-4e37-b868-8df174691786 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.273s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.364480] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12673a6-2ae5-4deb-bfdb-88ed6dc9a800 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.371274] env[65107]: DEBUG nova.network.neutron [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Successfully updated port: 8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 676.381448] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65f32a8-0c69-490a-941d-73c62d5c64d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.421741] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8c10cc-1be5-4489-bd28-14d61412824c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.431740] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dde65f-96b5-49f0-8c6b-44f041ca2edb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.447250] env[65107]: DEBUG nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 676.450976] env[65107]: DEBUG nova.compute.provider_tree [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.560152] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.560571] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.875419] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.876300] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.876676] env[65107]: DEBUG nova.network.neutron [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 676.963920] env[65107]: DEBUG nova.scheduler.client.report [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 677.146589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "7b24c530-56bf-4666-96f9-e2dc66728f1a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.146589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.146589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "7b24c530-56bf-4666-96f9-e2dc66728f1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.146589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.147022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.150814] env[65107]: INFO nova.compute.manager [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Terminating instance [ 677.348300] env[65107]: WARNING openstack [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.348300] env[65107]: WARNING openstack [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.376967] env[65107]: WARNING neutronclient.v2_0.client [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.377730] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.378083] env[65107]: WARNING openstack [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.386178] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.386532] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.472656] env[65107]: DEBUG nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 677.517703] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 677.519126] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 677.519126] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 677.519126] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 677.519126] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 677.519416] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 677.519831] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.520097] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 677.520385] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 677.520611] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 677.520923] env[65107]: DEBUG nova.virt.hardware [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 677.522359] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a09c1c8-f2fb-4503-b472-dddde6961ea7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.537481] env[65107]: DEBUG nova.network.neutron [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 677.541765] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf6a3bf-6e6c-45d4-8760-f1356c77509c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.574341] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 677.577856] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Creating folder: Project (ba25b2abe65f41a98d8db49cbc2a1e2d). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 677.578276] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5426ee80-9895-4f40-912c-eafc06789937 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.590962] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Created folder: Project (ba25b2abe65f41a98d8db49cbc2a1e2d) in parent group-v992574. [ 677.591182] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Creating folder: Instances. Parent ref: group-v992647. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 677.591539] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75673e30-56ad-4e79-8e5f-a81d7f1cb015 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.605316] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Created folder: Instances in parent group-v992647. [ 677.605725] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 677.607616] env[65107]: WARNING neutronclient.v2_0.client [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.608463] env[65107]: WARNING openstack [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.609060] env[65107]: WARNING openstack [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.620138] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 677.620138] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba1b7974-534d-4373-b210-dc9786779f74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.639022] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.639022] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.650581] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 677.650581] env[65107]: value = "task-5102400" [ 677.650581] env[65107]: _type = "Task" [ 677.650581] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.654832] env[65107]: DEBUG nova.compute.manager [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 677.655107] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.656460] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a5db86-7492-40f1-a10b-a934d7b3c16c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.669186] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102400, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.671011] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 677.675024] env[65107]: DEBUG nova.network.neutron [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updated VIF entry in instance network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 677.675024] env[65107]: DEBUG nova.network.neutron [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.675679] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4dd7242b-25de-433e-b911-7269f1a794c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.689368] env[65107]: DEBUG oslo_vmware.api [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 677.689368] env[65107]: value = "task-5102401" [ 677.689368] env[65107]: _type = "Task" [ 677.689368] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.700956] env[65107]: DEBUG oslo_vmware.api [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.768981] env[65107]: DEBUG nova.compute.manager [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Received event network-changed-8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 677.769399] env[65107]: DEBUG nova.compute.manager [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Refreshing instance network info cache due to event network-changed-8226470b-4e4b-4b3f-b40f-b29d3149af69. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 677.769552] env[65107]: DEBUG oslo_concurrency.lockutils [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Acquiring lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.769704] env[65107]: DEBUG oslo_concurrency.lockutils [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Acquired lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.769872] env[65107]: DEBUG nova.network.neutron [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Refreshing network info cache for port 8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 677.881512] env[65107]: DEBUG nova.network.neutron [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updated VIF entry in instance network info cache for port 02108638-8e1e-447a-9b94-a45adfd32161. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 677.881913] env[65107]: DEBUG nova.network.neutron [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updating instance_info_cache with network_info: [{"id": "02108638-8e1e-447a-9b94-a45adfd32161", "address": "fa:16:3e:49:92:f3", "network": {"id": "f011d711-ac42-4e0b-bcd5-8f569242741c", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-475912691-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ac04479a6164cf7b18b9ad304a0c9b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24727047-6358-4015-86c1-394ab07fb88f", "external-id": "nsx-vlan-transportzone-476", "segmentation_id": 476, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02108638-8e", "ovs_interfaceid": "02108638-8e1e-447a-9b94-a45adfd32161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.944811] env[65107]: WARNING neutronclient.v2_0.client [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.945962] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.946988] env[65107]: WARNING openstack [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.989524] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.556s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.993980] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.730s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.994309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.994549] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 677.994911] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.505s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.996412] env[65107]: INFO nova.compute.claims [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.999783] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f21e369-6165-4ad8-b7f9-7e2a8f9cfd9c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.010865] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2205285b-6989-47ed-bac4-5d6efc16b1ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.027301] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df333d0-50e3-4a7e-a216-28cd01a387ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.035931] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962e6007-aa2b-4970-879c-fd64e5173706 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.071132] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178912MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 678.071132] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.136964] env[65107]: DEBUG nova.network.neutron [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Updating instance_info_cache with network_info: [{"id": "8db11d31-614c-4e71-bb14-c5327b541437", "address": "fa:16:3e:46:96:e2", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8db11d31-61", "ovs_interfaceid": "8db11d31-614c-4e71-bb14-c5327b541437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 678.167697] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102400, 'name': CreateVM_Task, 'duration_secs': 0.378348} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.167994] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 678.168309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.168467] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.168777] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 678.169044] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b1808e6-c72e-4962-aa5b-33946ffaad91 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.176791] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 678.176791] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4820e-5e37-171d-d678-6eb5b48c2ac5" [ 678.176791] env[65107]: _type = "Task" [ 678.176791] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.185258] env[65107]: DEBUG oslo_concurrency.lockutils [req-e95f3a39-c9b0-4394-9d8a-336af0e23280 req-ebdc11b4-05e4-49d3-8f97-a1cb7763ddcb service nova] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.194550] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4820e-5e37-171d-d678-6eb5b48c2ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.205855] env[65107]: DEBUG oslo_vmware.api [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102401, 'name': PowerOffVM_Task, 'duration_secs': 0.282914} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.206200] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 678.207247] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 678.207247] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e91e8b35-b5af-4c79-bf01-27802288f176 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.276819] env[65107]: WARNING neutronclient.v2_0.client [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.277154] env[65107]: WARNING openstack [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.277792] env[65107]: WARNING openstack [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.287972] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 678.289473] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 678.289473] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Deleting the datastore file [datastore1] 7b24c530-56bf-4666-96f9-e2dc66728f1a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 678.289473] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7129ceac-22b4-4219-9ef5-424a1063f151 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.300953] env[65107]: DEBUG oslo_vmware.api [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for the task: (returnval){ [ 678.300953] env[65107]: value = "task-5102403" [ 678.300953] env[65107]: _type = "Task" [ 678.300953] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.309996] env[65107]: DEBUG oslo_vmware.api [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102403, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.389315] env[65107]: DEBUG oslo_concurrency.lockutils [req-2168fb80-d97e-46bd-8209-b2adeff19287 req-fb81aa25-dc4a-4700-9df4-8ca778e98175 service nova] Releasing lock "refresh_cache-ccd68268-d0fc-406e-896f-c61b114dd75d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.472115] env[65107]: WARNING openstack [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.472115] env[65107]: WARNING openstack [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.565231] env[65107]: WARNING neutronclient.v2_0.client [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.565231] env[65107]: WARNING openstack [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.565231] env[65107]: WARNING openstack [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.588320] env[65107]: INFO nova.scheduler.client.report [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleted allocation for migration 6662c375-2493-41ad-b464-55f88c6fa7c1 [ 678.638193] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.638973] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Instance network_info: |[{"id": "8db11d31-614c-4e71-bb14-c5327b541437", "address": "fa:16:3e:46:96:e2", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8db11d31-61", "ovs_interfaceid": "8db11d31-614c-4e71-bb14-c5327b541437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 678.639592] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:96:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8db11d31-614c-4e71-bb14-c5327b541437', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.652425] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 678.654056] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad53146e-210d-4321-b819-90795c1af260] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 678.654056] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b83f1b3-381f-46f8-b251-3839b2b5bef3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.698135] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4820e-5e37-171d-d678-6eb5b48c2ac5, 'name': SearchDatastore_Task, 'duration_secs': 0.026458} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.699705] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.699961] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 678.700342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.700506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.700730] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 678.701095] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.701095] env[65107]: value = "task-5102404" [ 678.701095] env[65107]: _type = "Task" [ 678.701095] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.701317] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d8854a7-4ff1-4d7e-9584-850fa5d15383 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.716540] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102404, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.717993] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 678.718190] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 678.719081] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc36c511-5eaa-4aea-91da-1f10198ec992 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.727878] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 678.727878] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5240d461-7651-3bca-aa15-8294d395cef9" [ 678.727878] env[65107]: _type = "Task" [ 678.727878] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.741358] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5240d461-7651-3bca-aa15-8294d395cef9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.744191] env[65107]: DEBUG nova.network.neutron [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Updated VIF entry in instance network info cache for port 8226470b-4e4b-4b3f-b40f-b29d3149af69. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 678.744469] env[65107]: DEBUG nova.network.neutron [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Updating instance_info_cache with network_info: [{"id": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "address": "fa:16:3e:92:8d:9e", "network": {"id": "b9e7c478-0189-4385-98a5-bc3ca80526e5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1623056907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c5372cbba21404eb54eda5f08f0ac42", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea6e81c3-94aa-40a6-a4d4-7f338b503442", "external-id": "nsx-vlan-transportzone-637", "segmentation_id": 637, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8226470b-4e", "ovs_interfaceid": "8226470b-4e4b-4b3f-b40f-b29d3149af69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 678.812664] env[65107]: DEBUG oslo_vmware.api [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Task: {'id': task-5102403, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.449325} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.812926] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 678.813740] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 678.813740] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.813740] env[65107]: INFO nova.compute.manager [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 678.813740] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 678.814116] env[65107]: DEBUG nova.compute.manager [-] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 678.814116] env[65107]: DEBUG nova.network.neutron [-] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 678.814388] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.814947] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.815234] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.856613] env[65107]: DEBUG nova.compute.manager [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Received event network-vif-plugged-8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 678.856750] env[65107]: DEBUG oslo_concurrency.lockutils [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Acquiring lock "ad53146e-210d-4321-b819-90795c1af260-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.856944] env[65107]: DEBUG oslo_concurrency.lockutils [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Lock "ad53146e-210d-4321-b819-90795c1af260-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.856996] env[65107]: DEBUG oslo_concurrency.lockutils [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Lock "ad53146e-210d-4321-b819-90795c1af260-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.857182] env[65107]: DEBUG nova.compute.manager [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] No waiting events found dispatching network-vif-plugged-8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 678.857351] env[65107]: WARNING nova.compute.manager [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Received unexpected event network-vif-plugged-8db11d31-614c-4e71-bb14-c5327b541437 for instance with vm_state building and task_state spawning. [ 678.857518] env[65107]: DEBUG nova.compute.manager [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Received event network-changed-8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 678.857692] env[65107]: DEBUG nova.compute.manager [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Refreshing instance network info cache due to event network-changed-8db11d31-614c-4e71-bb14-c5327b541437. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 678.857897] env[65107]: DEBUG oslo_concurrency.lockutils [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Acquiring lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.858107] env[65107]: DEBUG oslo_concurrency.lockutils [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Acquired lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.858203] env[65107]: DEBUG nova.network.neutron [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Refreshing network info cache for port 8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 678.863795] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.864094] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.909702] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.100383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-796b2230-4abd-41d7-88d1-b803f4afcbfc tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.769s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.216766] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102404, 'name': CreateVM_Task, 'duration_secs': 0.448425} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.216766] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad53146e-210d-4321-b819-90795c1af260] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 679.216766] env[65107]: WARNING neutronclient.v2_0.client [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.216766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.216766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.217380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 679.217380] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34829a9c-fd6a-4cbd-af26-2479956cd5ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.223790] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 679.223790] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522364cd-617d-44e2-2f61-e5e85af4f44b" [ 679.223790] env[65107]: _type = "Task" [ 679.223790] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.241901] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522364cd-617d-44e2-2f61-e5e85af4f44b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.249251] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5240d461-7651-3bca-aa15-8294d395cef9, 'name': SearchDatastore_Task, 'duration_secs': 0.013708} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.249721] env[65107]: DEBUG oslo_concurrency.lockutils [req-53c6d19b-5d4e-47ed-9822-0094338282d1 req-6550e3a8-014d-4916-833a-86350334fc4d service nova] Releasing lock "refresh_cache-705b91ea-ccd2-41ce-a14f-2ea5b70f80df" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.253033] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-145d0b03-28fb-4ae6-b456-980d083e89f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.260986] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 679.260986] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5219c738-7944-a53e-41ea-ba6cf5aa08d5" [ 679.260986] env[65107]: _type = "Task" [ 679.260986] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.269761] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5219c738-7944-a53e-41ea-ba6cf5aa08d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.366362] env[65107]: WARNING neutronclient.v2_0.client [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.366362] env[65107]: WARNING openstack [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.366362] env[65107]: WARNING openstack [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.376420] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 679.499852] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2c7ede-cc17-4711-ac06-8c6de83d88ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.509784] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb97bce-ea33-4bda-bcdc-12a9dc164662 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.547814] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bf1dcb-c3c0-4e85-8c30-9d250c4166e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.558147] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341d27ae-e1a7-4737-812a-22abc3f13f3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.576031] env[65107]: DEBUG nova.compute.provider_tree [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.585777] env[65107]: WARNING openstack [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.586067] env[65107]: WARNING openstack [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.733524] env[65107]: WARNING neutronclient.v2_0.client [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.734622] env[65107]: WARNING openstack [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.736172] env[65107]: WARNING openstack [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.750915] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522364cd-617d-44e2-2f61-e5e85af4f44b, 'name': SearchDatastore_Task, 'duration_secs': 0.028613} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.753182] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.753182] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.753182] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.755528] env[65107]: DEBUG nova.network.neutron [-] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.778132] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5219c738-7944-a53e-41ea-ba6cf5aa08d5, 'name': SearchDatastore_Task, 'duration_secs': 0.027899} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.778588] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.778981] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] a2b72bac-6806-4b7b-b972-32aea52f7c82/a2b72bac-6806-4b7b-b972-32aea52f7c82.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 679.779785] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.781793] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.782433] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11323070-2850-4ad6-81a3-a26044fd0926 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.785949] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3705252d-31fd-4813-9b62-1af8a39b8f51 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.799053] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.799563] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 679.800447] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 679.800447] env[65107]: value = "task-5102406" [ 679.800447] env[65107]: _type = "Task" [ 679.800447] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.800447] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6c1de2e-b827-4284-9989-295a7d252d83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.814052] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.815186] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 679.815186] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52366c01-3d90-8da5-c7a1-d23c8dfb6d96" [ 679.815186] env[65107]: _type = "Task" [ 679.815186] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.825974] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 679.825974] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.833643] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52366c01-3d90-8da5-c7a1-d23c8dfb6d96, 'name': SearchDatastore_Task, 'duration_secs': 0.01176} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.834807] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c3309fc-58d3-4e85-8ef8-74f9d88c18c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.842233] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 679.842233] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528535e8-9580-8226-da98-483baba32c2e" [ 679.842233] env[65107]: _type = "Task" [ 679.842233] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.854439] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528535e8-9580-8226-da98-483baba32c2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.871390] env[65107]: DEBUG nova.network.neutron [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Updated VIF entry in instance network info cache for port 8db11d31-614c-4e71-bb14-c5327b541437. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 679.872639] env[65107]: DEBUG nova.network.neutron [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Updating instance_info_cache with network_info: [{"id": "8db11d31-614c-4e71-bb14-c5327b541437", "address": "fa:16:3e:46:96:e2", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8db11d31-61", "ovs_interfaceid": "8db11d31-614c-4e71-bb14-c5327b541437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.913712] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.082890] env[65107]: DEBUG nova.scheduler.client.report [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 680.259324] env[65107]: INFO nova.compute.manager [-] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Took 1.44 seconds to deallocate network for instance. [ 680.315286] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491082} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.315286] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] a2b72bac-6806-4b7b-b972-32aea52f7c82/a2b72bac-6806-4b7b-b972-32aea52f7c82.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 680.315465] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 680.316669] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e99edd89-5cb9-4104-a55b-1f6a777b577e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.324789] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 680.324789] env[65107]: value = "task-5102407" [ 680.324789] env[65107]: _type = "Task" [ 680.324789] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.328928] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 680.339488] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.360024] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528535e8-9580-8226-da98-483baba32c2e, 'name': SearchDatastore_Task, 'duration_secs': 0.012176} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.360396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.360690] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ad53146e-210d-4321-b819-90795c1af260/ad53146e-210d-4321-b819-90795c1af260.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 680.360987] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5946434e-ef41-4eda-bed9-4dfdad776b16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.373041] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 680.373041] env[65107]: value = "task-5102408" [ 680.373041] env[65107]: _type = "Task" [ 680.373041] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.379707] env[65107]: DEBUG oslo_concurrency.lockutils [req-3bc07848-a41b-40f5-ae5d-3f883a0c3123 req-ec4ef957-eb97-4194-bca4-1f7a1542fd2a service nova] Releasing lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.387206] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.589604] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.590501] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 680.594105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.926s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.596077] env[65107]: INFO nova.compute.claims [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.773047] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.839320] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082439} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.842769] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 680.844544] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0f12d2-c109-4c84-86c1-b59510a1b5a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.882733] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] a2b72bac-6806-4b7b-b972-32aea52f7c82/a2b72bac-6806-4b7b-b972-32aea52f7c82.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 680.884750] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.886954] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc461610-822d-4eee-8530-a2deac6c96cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.913277] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.913658] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.925397] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505671} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.925579] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 680.925579] env[65107]: value = "task-5102409" [ 680.925579] env[65107]: _type = "Task" [ 680.925579] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.925774] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ad53146e-210d-4321-b819-90795c1af260/ad53146e-210d-4321-b819-90795c1af260.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 680.926580] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 680.926580] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11ba1677-1ced-4ed8-b6f5-91f4e4fd9155 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.940895] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.944306] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 680.944306] env[65107]: value = "task-5102410" [ 680.944306] env[65107]: _type = "Task" [ 680.944306] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.954070] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102410, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.960586] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.960862] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.003649] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "90706f49-af87-4038-ae76-02a95e4a12e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.003933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "90706f49-af87-4038-ae76-02a95e4a12e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.102441] env[65107]: DEBUG nova.compute.utils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 681.105899] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 681.106642] env[65107]: DEBUG nova.network.neutron [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 681.107018] env[65107]: WARNING neutronclient.v2_0.client [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.107736] env[65107]: WARNING neutronclient.v2_0.client [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.108649] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.108962] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.188752] env[65107]: DEBUG nova.policy [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2493e92666b14fe1b13275903407fcc8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a4457697290e40e89d4261fe02ff5eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 681.417653] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 681.439906] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.452664] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102410, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107378} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.452955] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 681.454116] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a1f66b-82d6-44e4-816e-ad85cacc9eb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.474920] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 681.489548] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] ad53146e-210d-4321-b819-90795c1af260/ad53146e-210d-4321-b819-90795c1af260.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 681.492634] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e857929e-8bb3-4560-9c2e-33d090bc7673 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.516903] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 681.516903] env[65107]: value = "task-5102411" [ 681.516903] env[65107]: _type = "Task" [ 681.516903] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.527728] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102411, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.620835] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 681.636872] env[65107]: DEBUG nova.network.neutron [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Successfully created port: 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 681.942860] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102409, 'name': ReconfigVM_Task, 'duration_secs': 0.732816} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.942860] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Reconfigured VM instance instance-0000001a to attach disk [datastore1] a2b72bac-6806-4b7b-b972-32aea52f7c82/a2b72bac-6806-4b7b-b972-32aea52f7c82.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 681.943669] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18291517-6677-45ed-bc85-cee2ee2e4749 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.947169] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.955990] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 681.955990] env[65107]: value = "task-5102413" [ 681.955990] env[65107]: _type = "Task" [ 681.955990] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.966975] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102413, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.998332] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.030706] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102411, 'name': ReconfigVM_Task, 'duration_secs': 0.321517} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.031036] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Reconfigured VM instance instance-00000019 to attach disk [datastore1] ad53146e-210d-4321-b819-90795c1af260/ad53146e-210d-4321-b819-90795c1af260.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 682.031637] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b09fa35-37f1-4f08-8eb9-4fa1383ae419 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.041785] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 682.041785] env[65107]: value = "task-5102414" [ 682.041785] env[65107]: _type = "Task" [ 682.041785] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.055644] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102414, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.115822] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babc74be-4dec-4995-84da-d05350594383 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.124645] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb500f0-e915-49f0-b481-42d0d6ace894 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.165039] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13c7757-f8c5-4840-890b-5197cbde42f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.173655] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550b3c56-d336-4740-9170-15aa8d40ecf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.189256] env[65107]: DEBUG nova.compute.provider_tree [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 682.275477] env[65107]: DEBUG nova.compute.manager [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 682.276506] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961e8e51-aff3-43f6-a52d-092621ab6818 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.466422] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102413, 'name': Rename_Task, 'duration_secs': 0.14765} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.466764] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 682.467043] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f390c2d7-3c0a-42e5-9194-f7f1118ff384 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.473159] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 682.473159] env[65107]: value = "task-5102415" [ 682.473159] env[65107]: _type = "Task" [ 682.473159] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.481599] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102415, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.554341] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102414, 'name': Rename_Task, 'duration_secs': 0.160291} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.554901] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 682.555325] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d766bfc-355c-468d-b60a-a71dabe03b33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.565329] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 682.565329] env[65107]: value = "task-5102416" [ 682.565329] env[65107]: _type = "Task" [ 682.565329] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.575518] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.633455] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 682.665953] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 682.666227] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 682.666385] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 682.666615] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 682.666793] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 682.666943] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 682.667170] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.667332] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 682.667498] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 682.667661] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 682.667832] env[65107]: DEBUG nova.virt.hardware [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 682.668787] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fdfd2f-1603-496a-b603-ff26a6159b21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.678992] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf737aab-db9d-4acf-b673-f18a5f26a64a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.721754] env[65107]: ERROR nova.scheduler.client.report [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [req-b2361da8-5b99-419c-a786-b05fd9459126] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b2361da8-5b99-419c-a786-b05fd9459126"}]} [ 682.743758] env[65107]: DEBUG nova.scheduler.client.report [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 682.765365] env[65107]: DEBUG nova.scheduler.client.report [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 682.765615] env[65107]: DEBUG nova.compute.provider_tree [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 682.783737] env[65107]: DEBUG nova.scheduler.client.report [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: 845eef68-4dc1-42f4-981b-f1d84ef23d25 {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 682.792760] env[65107]: INFO nova.compute.manager [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] instance snapshotting [ 682.795433] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e884752-f84f-41a9-bc5e-3bbf21853202 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.819509] env[65107]: DEBUG nova.scheduler.client.report [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 682.822532] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f218c1f8-c9b3-4e3b-828c-7e967ab9e924 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.992351] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102415, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.018740] env[65107]: DEBUG nova.compute.manager [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Received event network-changed-55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 683.018951] env[65107]: DEBUG nova.compute.manager [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Refreshing instance network info cache due to event network-changed-55102692-2502-4633-b00b-cc29760fb310. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 683.023127] env[65107]: DEBUG oslo_concurrency.lockutils [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Acquiring lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.023127] env[65107]: DEBUG oslo_concurrency.lockutils [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Acquired lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.023306] env[65107]: DEBUG nova.network.neutron [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Refreshing network info cache for port 55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 683.082542] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102416, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.338042] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 683.338042] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-22670497-51df-4874-8345-8c0f406ebd72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.346497] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 683.346497] env[65107]: value = "task-5102417" [ 683.346497] env[65107]: _type = "Task" [ 683.346497] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.356662] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102417, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.386941] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c247264-f444-4b88-a726-74d115c04230 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.401515] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cff6118-42fa-41c9-8f35-73c6c5dab74f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.444584] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a86cdeb-9f23-4953-8a9a-82524f4a7051 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.456245] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0896d7-16fc-4cfe-8d4a-8712a5b9351e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.463670] env[65107]: DEBUG nova.network.neutron [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Successfully updated port: 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 683.474315] env[65107]: DEBUG nova.compute.provider_tree [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 683.487311] env[65107]: DEBUG oslo_vmware.api [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102415, 'name': PowerOnVM_Task, 'duration_secs': 0.566636} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.487753] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 683.489326] env[65107]: INFO nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Took 6.02 seconds to spawn the instance on the hypervisor. [ 683.489326] env[65107]: DEBUG nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 683.489928] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762be51d-717d-4aa1-833a-621cda587bcc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.533648] env[65107]: WARNING neutronclient.v2_0.client [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.534516] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.534851] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.582236] env[65107]: DEBUG oslo_vmware.api [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102416, 'name': PowerOnVM_Task, 'duration_secs': 0.570332} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.584180] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 683.584485] env[65107]: INFO nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Took 8.29 seconds to spawn the instance on the hypervisor. [ 683.584667] env[65107]: DEBUG nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 683.585838] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91209928-f054-42cc-b33f-ab3fd93ad31d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.719449] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.719449] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.813020] env[65107]: WARNING neutronclient.v2_0.client [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.813907] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.814546] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.859948] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102417, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.977574] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.978053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.978349] env[65107]: DEBUG nova.network.neutron [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 683.986833] env[65107]: DEBUG nova.network.neutron [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Updated VIF entry in instance network info cache for port 55102692-2502-4633-b00b-cc29760fb310. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 683.987145] env[65107]: DEBUG nova.network.neutron [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Updating instance_info_cache with network_info: [{"id": "55102692-2502-4633-b00b-cc29760fb310", "address": "fa:16:3e:6a:99:59", "network": {"id": "90ceb5af-809c-4b48-9c35-26e48235753c", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-156477542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9fcd71f4f064b488f832406fedafbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55102692-25", "ovs_interfaceid": "55102692-2502-4633-b00b-cc29760fb310", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 684.012917] env[65107]: INFO nova.compute.manager [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Took 19.59 seconds to build instance. [ 684.032926] env[65107]: DEBUG nova.scheduler.client.report [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 44 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 684.033608] env[65107]: DEBUG nova.compute.provider_tree [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 44 to 45 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 684.033608] env[65107]: DEBUG nova.compute.provider_tree [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 684.111124] env[65107]: INFO nova.compute.manager [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Took 20.14 seconds to build instance. [ 684.358896] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102417, 'name': CreateSnapshot_Task, 'duration_secs': 0.732748} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.360145] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 684.360145] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1aa502e-f10d-4d35-bd02-a6ba6ca1e8a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.483113] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.483611] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.497080] env[65107]: DEBUG oslo_concurrency.lockutils [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Releasing lock "refresh_cache-0e6c9f2d-4412-4d8d-8348-6552a34ab349" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 684.497234] env[65107]: DEBUG nova.compute.manager [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Received event network-changed-09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 684.497411] env[65107]: DEBUG nova.compute.manager [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Refreshing instance network info cache due to event network-changed-09df7af1-1609-4648-a7d1-a870687d5bbb. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 684.497705] env[65107]: DEBUG oslo_concurrency.lockutils [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Acquiring lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.498951] env[65107]: DEBUG oslo_concurrency.lockutils [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Acquired lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 684.498951] env[65107]: DEBUG nova.network.neutron [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Refreshing network info cache for port 09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 684.514191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16ed3b5c-5445-4ed0-ba50-b7edec8d475a tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "a2b72bac-6806-4b7b-b972-32aea52f7c82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.117s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.529394] env[65107]: DEBUG nova.network.neutron [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 684.541065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.947s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.541789] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 684.544278] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.754s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.545035] env[65107]: DEBUG nova.objects.instance [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lazy-loading 'resources' on Instance uuid 7a3009bf-54a2-4565-a1aa-1d19286a4810 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.567117] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.567527] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.614451] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be6e114d-51ce-46a2-b50a-3a25984ee5d0 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "ad53146e-210d-4321-b819-90795c1af260" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.682s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.660120] env[65107]: WARNING neutronclient.v2_0.client [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.660603] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.660941] env[65107]: WARNING openstack [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.701430] env[65107]: DEBUG nova.compute.manager [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 684.702534] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898b7c4b-47a1-4087-ba86-78c9ea480a61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.832085] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "ccd68268-d0fc-406e-896f-c61b114dd75d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.832357] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.832564] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "ccd68268-d0fc-406e-896f-c61b114dd75d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.832790] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.832984] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.835898] env[65107]: INFO nova.compute.manager [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Terminating instance [ 684.883458] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 684.884770] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d958e1fc-46c0-4373-b77b-64cf6768b1a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.894735] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 684.894735] env[65107]: value = "task-5102419" [ 684.894735] env[65107]: _type = "Task" [ 684.894735] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.904414] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102419, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.907370] env[65107]: DEBUG nova.network.neutron [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 684.954946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.955183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 685.002368] env[65107]: WARNING neutronclient.v2_0.client [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.003091] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.003516] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.017054] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 685.048565] env[65107]: DEBUG nova.compute.utils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 685.054342] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 685.054653] env[65107]: DEBUG nova.network.neutron [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 685.055163] env[65107]: WARNING neutronclient.v2_0.client [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.055442] env[65107]: WARNING neutronclient.v2_0.client [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.056285] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.056541] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.212277] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "d40aec5a-d898-42f3-a121-503cf043f72b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.212502] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "d40aec5a-d898-42f3-a121-503cf043f72b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 685.214695] env[65107]: INFO nova.compute.manager [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] instance snapshotting [ 685.217324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9868879f-ada8-4105-b86e-cba9ec08615b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.231772] env[65107]: DEBUG nova.policy [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6590cb74e83e45bd944f23fa888a4b04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f379144b78764fe394039d87b043a946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 685.251731] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b768dbde-5f29-4c39-bf6b-631de670ecec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.346694] env[65107]: DEBUG nova.compute.manager [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 685.347266] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.349077] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0aac15b-83e3-4d8b-a7a1-f5fa7fc3327f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.363280] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 685.363724] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-919cdc3b-b44d-4fcd-bbc5-32e6e3669e7a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.376352] env[65107]: DEBUG oslo_vmware.api [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 685.376352] env[65107]: value = "task-5102420" [ 685.376352] env[65107]: _type = "Task" [ 685.376352] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.378826] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.378826] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.399223] env[65107]: DEBUG oslo_vmware.api [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.410015] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102419, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.413770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.414232] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Instance network_info: |[{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 685.414997] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:9f:84', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a15de394-0367-4921-a5c1-6ac8615e3283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41ad83e5-4645-4e91-a223-473eefb69c38', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 685.423579] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Creating folder: Project (a4457697290e40e89d4261fe02ff5eb7). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 685.429908] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-face0130-0e88-450b-94ad-c30d850de0b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.445395] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Created folder: Project (a4457697290e40e89d4261fe02ff5eb7) in parent group-v992574. [ 685.445626] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Creating folder: Instances. Parent ref: group-v992654. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 685.446051] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1dbd2e8c-f489-470c-8f40-893dc0216049 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.460883] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 685.465362] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Created folder: Instances in parent group-v992654. [ 685.465929] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 685.467729] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 685.468101] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2213e0f6-c560-4d19-81ac-7048324c4fc4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.494092] env[65107]: WARNING neutronclient.v2_0.client [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.496435] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.496435] env[65107]: WARNING openstack [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.511612] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 685.511612] env[65107]: value = "task-5102423" [ 685.511612] env[65107]: _type = "Task" [ 685.511612] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.527046] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102423, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.546483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.553310] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 685.683969] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246631a0-591b-4c65-b4e1-315ac45fd30b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.692903] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d458355-7209-4cbf-a598-b7ffe4b79e7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.726347] env[65107]: DEBUG nova.network.neutron [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Successfully created port: 3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 685.731032] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb3dae1-2d06-4637-8f89-0b5fb76565b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.741515] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819040c0-56d6-4f5c-97c5-84a8bf41dc19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.769672] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 685.769672] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 685.769672] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a8b6f955-01fd-47e2-a842-9c4277dbedd8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.777882] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 685.777882] env[65107]: value = "task-5102424" [ 685.777882] env[65107]: _type = "Task" [ 685.777882] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.780914] env[65107]: DEBUG nova.network.neutron [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updated VIF entry in instance network info cache for port 09df7af1-1609-4648-a7d1-a870687d5bbb. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 685.781577] env[65107]: DEBUG nova.network.neutron [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updating instance_info_cache with network_info: [{"id": "09df7af1-1609-4648-a7d1-a870687d5bbb", "address": "fa:16:3e:d9:57:dd", "network": {"id": "de1c1b5c-0632-45e3-9134-5fbc553a0cf0", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-842310743-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3fa2b26480f4495926fadf02489de3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09df7af1-16", "ovs_interfaceid": "09df7af1-1609-4648-a7d1-a870687d5bbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.794921] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102424, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.897180] env[65107]: DEBUG oslo_vmware.api [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102420, 'name': PowerOffVM_Task, 'duration_secs': 0.237121} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.897533] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 685.897703] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 685.898085] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-916ccf8c-8606-454e-a332-017be9c7b312 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.913144] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102419, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.989913] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 685.989913] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 685.989913] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Deleting the datastore file [datastore1] ccd68268-d0fc-406e-896f-c61b114dd75d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 685.989913] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cbcdde6a-181d-4653-b4c0-bed1397e94ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.999088] env[65107]: DEBUG oslo_vmware.api [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for the task: (returnval){ [ 685.999088] env[65107]: value = "task-5102427" [ 685.999088] env[65107]: _type = "Task" [ 685.999088] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.005822] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.016731] env[65107]: DEBUG oslo_vmware.api [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102427, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.028662] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102423, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.111351] env[65107]: DEBUG nova.compute.manager [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-vif-plugged-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 686.111608] env[65107]: DEBUG oslo_concurrency.lockutils [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Acquiring lock "b20a2c71-514d-488a-8edb-8fb0274a894a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.111906] env[65107]: DEBUG oslo_concurrency.lockutils [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.112112] env[65107]: DEBUG oslo_concurrency.lockutils [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 686.112301] env[65107]: DEBUG nova.compute.manager [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] No waiting events found dispatching network-vif-plugged-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 686.112971] env[65107]: WARNING nova.compute.manager [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received unexpected event network-vif-plugged-41ad83e5-4645-4e91-a223-473eefb69c38 for instance with vm_state building and task_state spawning. [ 686.112971] env[65107]: DEBUG nova.compute.manager [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 686.112971] env[65107]: DEBUG nova.compute.manager [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing instance network info cache due to event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 686.112971] env[65107]: DEBUG oslo_concurrency.lockutils [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.113370] env[65107]: DEBUG oslo_concurrency.lockutils [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.113370] env[65107]: DEBUG nova.network.neutron [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 686.289254] env[65107]: DEBUG oslo_concurrency.lockutils [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] Releasing lock "refresh_cache-7062b119-01a4-45f4-bb52-212178382271" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 686.289376] env[65107]: DEBUG nova.compute.manager [req-ca8cbb6e-411c-43b6-b8e3-29ad6fb8900c req-ece4a301-edcb-499d-889d-8a4d32a1e32a service nova] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Received event network-vif-deleted-d533dd9c-4bb6-4f96-b6fb-84bd5623a8ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 686.290560] env[65107]: DEBUG oslo_concurrency.lockutils [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Acquiring lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.290720] env[65107]: DEBUG oslo_concurrency.lockutils [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Acquired lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.290884] env[65107]: DEBUG nova.network.neutron [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] [instance: ad53146e-210d-4321-b819-90795c1af260] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 686.292800] env[65107]: ERROR nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [req-9a613de3-8d36-4fda-ba94-2aa26950b003] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9a613de3-8d36-4fda-ba94-2aa26950b003"}]} [ 686.300163] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102424, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.313208] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 686.335301] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 686.335538] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 686.355358] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 686.359713] env[65107]: DEBUG nova.compute.manager [None req-7e65a28b-b6b4-42f5-a38b-a5191ed6aecd tempest-ServerDiagnosticsV248Test-880666234 tempest-ServerDiagnosticsV248Test-880666234-project-admin] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 686.360987] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a2dde2-39b4-4aa0-904a-99be7c687f61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.370426] env[65107]: INFO nova.compute.manager [None req-7e65a28b-b6b4-42f5-a38b-a5191ed6aecd tempest-ServerDiagnosticsV248Test-880666234 tempest-ServerDiagnosticsV248Test-880666234-project-admin] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Retrieving diagnostics [ 686.373070] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7e4c54-5439-4ea9-9578-bbd29c460cb9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.412749] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 686.425632] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102419, 'name': CloneVM_Task, 'duration_secs': 1.354193} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.425892] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Created linked-clone VM from snapshot [ 686.428380] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a98748-6d51-4542-a235-5ca365e71d96 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.437562] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Uploading image 91f5fcd9-ad64-48e2-9af0-1f64e8a6cf13 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 686.460374] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 686.460374] env[65107]: value = "vm-992653" [ 686.460374] env[65107]: _type = "VirtualMachine" [ 686.460374] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 686.460677] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1fd9830d-53f7-4223-a190-eded1fa5a9fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.471591] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lease: (returnval){ [ 686.471591] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae061a-d413-ae8e-a52f-58d368c2579b" [ 686.471591] env[65107]: _type = "HttpNfcLease" [ 686.471591] env[65107]: } obtained for exporting VM: (result){ [ 686.471591] env[65107]: value = "vm-992653" [ 686.471591] env[65107]: _type = "VirtualMachine" [ 686.471591] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 686.471919] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the lease: (returnval){ [ 686.471919] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae061a-d413-ae8e-a52f-58d368c2579b" [ 686.471919] env[65107]: _type = "HttpNfcLease" [ 686.471919] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 686.481057] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 686.481057] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae061a-d413-ae8e-a52f-58d368c2579b" [ 686.481057] env[65107]: _type = "HttpNfcLease" [ 686.481057] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 686.512039] env[65107]: DEBUG oslo_vmware.api [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Task: {'id': task-5102427, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25827} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.512330] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 686.512466] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 686.512643] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.512816] env[65107]: INFO nova.compute.manager [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 686.513066] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 686.513257] env[65107]: DEBUG nova.compute.manager [-] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 686.513351] env[65107]: DEBUG nova.network.neutron [-] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 686.513605] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.514144] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.514397] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.531624] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102423, 'name': CreateVM_Task, 'duration_secs': 0.579227} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.531801] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 686.532311] env[65107]: WARNING neutronclient.v2_0.client [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.532670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.532842] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.533151] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 686.535763] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7495500f-6b1b-49c0-9bed-f0dbf3cd560a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.541906] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 686.541906] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52570e52-17e8-0251-df49-cb6f466c0284" [ 686.541906] env[65107]: _type = "Task" [ 686.541906] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.551782] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52570e52-17e8-0251-df49-cb6f466c0284, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.566224] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 686.588770] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 686.589015] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 686.589178] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 686.589405] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 686.589496] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 686.589651] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 686.589985] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.590074] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 686.590200] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 686.590361] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 686.590532] env[65107]: DEBUG nova.virt.hardware [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 686.591453] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120eb07e-9181-420d-8dd1-5e601aa670ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.595217] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.606046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e377b9a9-b7d2-4b68-baf7-3b25b0851b00 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.622018] env[65107]: WARNING neutronclient.v2_0.client [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.622683] env[65107]: WARNING openstack [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.623039] env[65107]: WARNING openstack [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.798154] env[65107]: WARNING neutronclient.v2_0.client [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.798790] env[65107]: WARNING openstack [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.798903] env[65107]: WARNING openstack [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.806964] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102424, 'name': CreateSnapshot_Task, 'duration_secs': 0.605018} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.807474] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 686.808471] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c10cc5f-70f1-4628-8045-9f369fd411d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.947217] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91d4115-8675-4991-b8da-33eebefa9c2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.956047] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d755fa4-188e-4eb0-bc93-61e1df18b1d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.993313] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd780dce-af9b-4d9c-ac15-d301d8827524 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.003169] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 687.003169] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae061a-d413-ae8e-a52f-58d368c2579b" [ 687.003169] env[65107]: _type = "HttpNfcLease" [ 687.003169] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 687.006128] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 687.006128] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae061a-d413-ae8e-a52f-58d368c2579b" [ 687.006128] env[65107]: _type = "HttpNfcLease" [ 687.006128] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 687.006446] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c7fb59-b294-41eb-af5a-d9f03540df72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.011393] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17044724-87c0-4183-89e0-64d6cd883d9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.032724] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e10295-7adf-db1c-f6a4-ade829428807/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 687.033935] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e10295-7adf-db1c-f6a4-ade829428807/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 687.035593] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 687.115284] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52570e52-17e8-0251-df49-cb6f466c0284, 'name': SearchDatastore_Task, 'duration_secs': 0.028566} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.116025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 687.116025] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 687.116176] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.116391] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.116471] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 687.116693] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-291b7bec-b2af-4964-9aac-670b46f4a847 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.128091] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 687.128307] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 687.129091] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03d20914-2619-425c-8dbc-27a6ad2c931e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.138713] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 687.138713] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5260f9fa-18bf-b0df-be04-af13d424b7d3" [ 687.138713] env[65107]: _type = "Task" [ 687.138713] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.145142] env[65107]: WARNING openstack [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.145142] env[65107]: WARNING openstack [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.158962] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5260f9fa-18bf-b0df-be04-af13d424b7d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.162158] env[65107]: WARNING openstack [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.162158] env[65107]: WARNING openstack [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.170840] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9ecaa5c0-062b-4685-9380-c3d72b5339dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.249496] env[65107]: WARNING neutronclient.v2_0.client [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.250354] env[65107]: WARNING openstack [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.250740] env[65107]: WARNING openstack [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.286577] env[65107]: WARNING neutronclient.v2_0.client [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.287301] env[65107]: WARNING openstack [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.287645] env[65107]: WARNING openstack [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.303687] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.303882] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.336763] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 687.344754] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b80615af-7ce2-4e5f-85ee-f4f3f7e23482 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.354711] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 687.354711] env[65107]: value = "task-5102429" [ 687.354711] env[65107]: _type = "Task" [ 687.354711] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.365060] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102429, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.383224] env[65107]: DEBUG nova.network.neutron [-] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.391391] env[65107]: DEBUG nova.compute.manager [req-9e844177-e395-40ba-9455-d30cd7e3e63d req-2256fef7-dda5-4214-b635-ea2ef6a93932 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Received event network-vif-deleted-02108638-8e1e-447a-9b94-a45adfd32161 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 687.391391] env[65107]: INFO nova.compute.manager [req-9e844177-e395-40ba-9455-d30cd7e3e63d req-2256fef7-dda5-4214-b635-ea2ef6a93932 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Neutron deleted interface 02108638-8e1e-447a-9b94-a45adfd32161; detaching it from the instance and deleting it from the info cache [ 687.391391] env[65107]: DEBUG nova.network.neutron [req-9e844177-e395-40ba-9455-d30cd7e3e63d req-2256fef7-dda5-4214-b635-ea2ef6a93932 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.482042] env[65107]: DEBUG nova.network.neutron [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] [instance: ad53146e-210d-4321-b819-90795c1af260] Updating instance_info_cache with network_info: [{"id": "8db11d31-614c-4e71-bb14-c5327b541437", "address": "fa:16:3e:46:96:e2", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8db11d31-61", "ovs_interfaceid": "8db11d31-614c-4e71-bb14-c5327b541437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.519867] env[65107]: DEBUG nova.network.neutron [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updated VIF entry in instance network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 687.520416] env[65107]: DEBUG nova.network.neutron [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.548181] env[65107]: DEBUG nova.network.neutron [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Successfully updated port: 3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 687.577165] env[65107]: ERROR nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] [req-80286f92-5e7f-4cd7-bc6e-b3f2396a832a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-80286f92-5e7f-4cd7-bc6e-b3f2396a832a"}]} [ 687.600082] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 687.619160] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 687.619462] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 687.637500] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 687.650732] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5260f9fa-18bf-b0df-be04-af13d424b7d3, 'name': SearchDatastore_Task, 'duration_secs': 0.019273} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.651690] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eede3f7f-cc0b-4c86-9e23-9168849cc369 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.658612] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 687.658612] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523430c9-6882-eff9-cf23-0649e68b2c44" [ 687.658612] env[65107]: _type = "Task" [ 687.658612] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.668696] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523430c9-6882-eff9-cf23-0649e68b2c44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.669977] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 687.870938] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102429, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.892768] env[65107]: INFO nova.compute.manager [-] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Took 1.38 seconds to deallocate network for instance. [ 687.902549] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7bf1be0-b917-416a-967a-fc2cda8aa3cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.919465] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65616965-9c9c-4596-867e-476d488d70f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.974472] env[65107]: DEBUG nova.compute.manager [req-9e844177-e395-40ba-9455-d30cd7e3e63d req-2256fef7-dda5-4214-b635-ea2ef6a93932 service nova] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Detach interface failed, port_id=02108638-8e1e-447a-9b94-a45adfd32161, reason: Instance ccd68268-d0fc-406e-896f-c61b114dd75d could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 688.004947] env[65107]: DEBUG oslo_concurrency.lockutils [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Releasing lock "refresh_cache-ad53146e-210d-4321-b819-90795c1af260" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.004947] env[65107]: DEBUG nova.compute.manager [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] [instance: ad53146e-210d-4321-b819-90795c1af260] Inject network info {{(pid=65107) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 688.005980] env[65107]: DEBUG nova.compute.manager [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] [instance: ad53146e-210d-4321-b819-90795c1af260] network_info to inject: |[{"id": "8db11d31-614c-4e71-bb14-c5327b541437", "address": "fa:16:3e:46:96:e2", "network": {"id": "53dee04c-dde9-43a9-b4e7-f1336bc7235f", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1082441117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2d4844feda0432ebe1ef160b0070b83", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8db11d31-61", "ovs_interfaceid": "8db11d31-614c-4e71-bb14-c5327b541437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 688.005980] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] [instance: ad53146e-210d-4321-b819-90795c1af260] Reconfiguring VM instance to set the machine id {{(pid=65107) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 688.006585] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-992dc658-e7fc-405f-84e2-99846c8d4e21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.029577] env[65107]: DEBUG oslo_vmware.api [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Waiting for the task: (returnval){ [ 688.029577] env[65107]: value = "task-5102430" [ 688.029577] env[65107]: _type = "Task" [ 688.029577] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.029577] env[65107]: DEBUG oslo_concurrency.lockutils [req-beddfac7-1651-421b-82c9-dc99b49425fd req-106326c2-d1e4-4c6c-9322-7029a5c04838 service nova] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.043499] env[65107]: DEBUG oslo_vmware.api [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Task: {'id': task-5102430, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.051830] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.052158] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 688.052419] env[65107]: DEBUG nova.network.neutron [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 688.176302] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523430c9-6882-eff9-cf23-0649e68b2c44, 'name': SearchDatastore_Task, 'duration_secs': 0.022491} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.180087] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.183187] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/b20a2c71-514d-488a-8edb-8fb0274a894a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 688.183187] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3123ff27-ce1f-46e2-9d33-a6d12a43e130 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.190934] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 688.190934] env[65107]: value = "task-5102431" [ 688.190934] env[65107]: _type = "Task" [ 688.190934] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.202266] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.370206] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102429, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.407176] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 688.409132] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c5b062-4703-4fe4-aedd-120c93bec638 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.417279] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d90bfe2-9341-4b7a-9109-372dd938c6fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.453610] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12f1e7c-435f-4a7c-8acd-07196d64c965 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.467136] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116856e8-253c-4a94-88f4-40c9801bfe6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.487832] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 688.533319] env[65107]: DEBUG oslo_vmware.api [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] Task: {'id': task-5102430, 'name': ReconfigVM_Task, 'duration_secs': 0.258876} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.533738] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-78ffa79d-a114-4f40-ab34-7da96b916625 tempest-ServersAdminTestJSON-782441049 tempest-ServersAdminTestJSON-782441049-project-admin] [instance: ad53146e-210d-4321-b819-90795c1af260] Reconfigured VM instance to set the machine id {{(pid=65107) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 688.559814] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.559814] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.616681] env[65107]: DEBUG nova.compute.manager [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Received event network-vif-plugged-3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 688.617111] env[65107]: DEBUG oslo_concurrency.lockutils [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 688.617585] env[65107]: DEBUG oslo_concurrency.lockutils [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 688.617948] env[65107]: DEBUG oslo_concurrency.lockutils [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.619699] env[65107]: DEBUG nova.compute.manager [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] No waiting events found dispatching network-vif-plugged-3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 688.619699] env[65107]: WARNING nova.compute.manager [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Received unexpected event network-vif-plugged-3d38d576-1871-49a5-bc10-aef1a37abfa0 for instance with vm_state building and task_state spawning. [ 688.619699] env[65107]: DEBUG nova.compute.manager [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Received event network-changed-3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 688.619699] env[65107]: DEBUG nova.compute.manager [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Refreshing instance network info cache due to event network-changed-3d38d576-1871-49a5-bc10-aef1a37abfa0. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 688.619699] env[65107]: DEBUG oslo_concurrency.lockutils [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.629457] env[65107]: DEBUG nova.network.neutron [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 688.652926] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.652926] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.713478] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102431, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.762158] env[65107]: WARNING neutronclient.v2_0.client [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.763105] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.763806] env[65107]: WARNING openstack [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.882042] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102429, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.939974] env[65107]: DEBUG nova.network.neutron [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.041823] env[65107]: DEBUG nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 49 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 689.044306] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 49 to 50 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 689.044306] env[65107]: DEBUG nova.compute.provider_tree [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 689.202911] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645641} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.203300] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/b20a2c71-514d-488a-8edb-8fb0274a894a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 689.203734] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 689.204064] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbb3f931-1538-43a0-808a-d6d55571c4e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.212868] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 689.212868] env[65107]: value = "task-5102432" [ 689.212868] env[65107]: _type = "Task" [ 689.212868] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.223771] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102432, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.377642] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102429, 'name': CloneVM_Task, 'duration_secs': 1.677089} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.378018] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Created linked-clone VM from snapshot [ 689.378948] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32e873b-feba-4d9f-936e-39eca5eb6067 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.389926] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Uploading image c57c1202-2e52-423c-9825-ad7cbbb94e56 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 689.414933] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 689.414933] env[65107]: value = "vm-992658" [ 689.414933] env[65107]: _type = "VirtualMachine" [ 689.414933] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 689.415363] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-084897f4-d840-41d0-a209-c9008324e584 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.424539] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lease: (returnval){ [ 689.424539] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a0726-f769-aebe-4fc8-399991e86582" [ 689.424539] env[65107]: _type = "HttpNfcLease" [ 689.424539] env[65107]: } obtained for exporting VM: (result){ [ 689.424539] env[65107]: value = "vm-992658" [ 689.424539] env[65107]: _type = "VirtualMachine" [ 689.424539] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 689.425050] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the lease: (returnval){ [ 689.425050] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a0726-f769-aebe-4fc8-399991e86582" [ 689.425050] env[65107]: _type = "HttpNfcLease" [ 689.425050] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 689.434358] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 689.434358] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a0726-f769-aebe-4fc8-399991e86582" [ 689.434358] env[65107]: _type = "HttpNfcLease" [ 689.434358] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 689.445549] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.445986] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Instance network_info: |[{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 689.446357] env[65107]: DEBUG oslo_concurrency.lockutils [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.446545] env[65107]: DEBUG nova.network.neutron [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Refreshing network info cache for port 3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 689.448662] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:b8:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd903c404-a23a-40c0-a217-96d4bb2e5b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d38d576-1871-49a5-bc10-aef1a37abfa0', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.456897] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating folder: Project (f379144b78764fe394039d87b043a946). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 689.457923] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8fe2605d-670c-4166-a481-a48394a588ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.472186] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created folder: Project (f379144b78764fe394039d87b043a946) in parent group-v992574. [ 689.472570] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating folder: Instances. Parent ref: group-v992659. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 689.472933] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b4a447b-717c-429c-ae6a-85f94a59a431 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.486505] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created folder: Instances in parent group-v992659. [ 689.486813] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 689.487103] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 689.487369] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e649682-e383-42b6-901b-d94c41f556a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.511649] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.511649] env[65107]: value = "task-5102436" [ 689.511649] env[65107]: _type = "Task" [ 689.511649] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.524885] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102436, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.548897] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 5.004s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.552191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.278s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.554495] env[65107]: INFO nova.compute.claims [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.609822] env[65107]: INFO nova.scheduler.client.report [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Deleted allocations for instance 7a3009bf-54a2-4565-a1aa-1d19286a4810 [ 689.731990] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102432, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235462} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.731990] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 689.732515] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf4ab31-4693-4460-ada3-95236bda473a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.769646] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/b20a2c71-514d-488a-8edb-8fb0274a894a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 689.770395] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65811f11-3f7f-4781-8a49-6d2db4ce0c3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.793242] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 689.793242] env[65107]: value = "task-5102437" [ 689.793242] env[65107]: _type = "Task" [ 689.793242] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.804229] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102437, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.941120] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 689.941120] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a0726-f769-aebe-4fc8-399991e86582" [ 689.941120] env[65107]: _type = "HttpNfcLease" [ 689.941120] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 689.941772] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 689.941772] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a0726-f769-aebe-4fc8-399991e86582" [ 689.941772] env[65107]: _type = "HttpNfcLease" [ 689.941772] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 689.942688] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39737417-734a-4e72-862e-983da2fcd6ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.949318] env[65107]: WARNING neutronclient.v2_0.client [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.950091] env[65107]: WARNING openstack [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.950545] env[65107]: WARNING openstack [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.964684] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4c4ec-76d7-7d3d-addb-65be64c7f8e4/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 689.964885] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4c4ec-76d7-7d3d-addb-65be64c7f8e4/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 690.042493] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102436, 'name': CreateVM_Task, 'duration_secs': 0.449865} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.042493] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 690.042493] env[65107]: WARNING neutronclient.v2_0.client [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.043251] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.043329] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.043819] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 690.044642] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abcc32e6-0d9e-439c-ac18-0a9ba69142b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.051076] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 690.051076] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52781857-9f36-74a4-7078-5ba8918add12" [ 690.051076] env[65107]: _type = "Task" [ 690.051076] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.067162] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52781857-9f36-74a4-7078-5ba8918add12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.105542] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-27586fcc-0f17-483c-948d-df0621d8603e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.126249] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89be55af-4ccb-49e3-bf86-ccc7b9529fec tempest-TenantUsagesTestJSON-55507538 tempest-TenantUsagesTestJSON-55507538-project-member] Lock "7a3009bf-54a2-4565-a1aa-1d19286a4810" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.180s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.161301] env[65107]: WARNING openstack [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.161704] env[65107]: WARNING openstack [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.248805] env[65107]: WARNING neutronclient.v2_0.client [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.249484] env[65107]: WARNING openstack [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.249829] env[65107]: WARNING openstack [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.305347] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102437, 'name': ReconfigVM_Task, 'duration_secs': 0.421483} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.305675] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Reconfigured VM instance instance-0000001b to attach disk [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/b20a2c71-514d-488a-8edb-8fb0274a894a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 690.306395] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a38568a6-07b6-4dbb-a6e1-71a70970b6d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.315811] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 690.315811] env[65107]: value = "task-5102438" [ 690.315811] env[65107]: _type = "Task" [ 690.315811] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.327725] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102438, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.391312] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "369caf24-48a6-4640-8601-e30fd47453da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.391588] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.429541] env[65107]: DEBUG nova.network.neutron [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updated VIF entry in instance network info cache for port 3d38d576-1871-49a5-bc10-aef1a37abfa0. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 690.429907] env[65107]: DEBUG nova.network.neutron [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 690.564074] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52781857-9f36-74a4-7078-5ba8918add12, 'name': SearchDatastore_Task, 'duration_secs': 0.015216} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.564074] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 690.564074] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.564074] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.564517] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.564517] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.566307] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb69ab6e-87d5-43d6-ad4e-60648a51e7c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.577812] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.577812] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 690.578636] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fd63389-741f-48c7-92b7-f26b776938ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.585390] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 690.585390] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52327a06-e14f-67f6-373b-8630233315bb" [ 690.585390] env[65107]: _type = "Task" [ 690.585390] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.596375] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52327a06-e14f-67f6-373b-8630233315bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.695022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.695457] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.834038] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102438, 'name': Rename_Task, 'duration_secs': 0.213072} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.838268] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 690.839172] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3de55aa2-7ec8-4ec9-b71f-0de7c072b46b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.849131] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 690.849131] env[65107]: value = "task-5102439" [ 690.849131] env[65107]: _type = "Task" [ 690.849131] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.864959] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.935237] env[65107]: DEBUG oslo_concurrency.lockutils [req-5257a923-0534-47ae-9bd9-d2fc831fd8e9 req-040eb710-04db-4f41-95c0-ea92a8ed6079 service nova] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.100726] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52327a06-e14f-67f6-373b-8630233315bb, 'name': SearchDatastore_Task, 'duration_secs': 0.016046} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.105035] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8eae39d-fc27-450d-bdf1-a11356642806 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.112270] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 691.112270] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7e88a-91db-cb9b-7840-198654dbc277" [ 691.112270] env[65107]: _type = "Task" [ 691.112270] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.127182] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7e88a-91db-cb9b-7840-198654dbc277, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.283059] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857c54e5-5334-40c5-9171-fe78371680e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.294637] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf172ae-b6fd-4b01-af57-37fc44aaad2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.331166] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32dd88e-d0e2-4fb2-9ea1-a054ea95c7de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.342650] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17b7fb3-082f-42a7-9fed-ebf40f087f81 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.361272] env[65107]: DEBUG nova.compute.provider_tree [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.372753] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102439, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.625976] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7e88a-91db-cb9b-7840-198654dbc277, 'name': SearchDatastore_Task, 'duration_secs': 0.017865} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.626410] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.627031] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96/71e6e280-50c3-4655-a6a3-40ea4c650d96.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 691.627031] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7fd3846-5d70-41ab-9111-702cfb9adbfc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.637094] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 691.637094] env[65107]: value = "task-5102440" [ 691.637094] env[65107]: _type = "Task" [ 691.637094] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.649317] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102440, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.862862] env[65107]: DEBUG oslo_vmware.api [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102439, 'name': PowerOnVM_Task, 'duration_secs': 0.796108} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.863263] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 691.864271] env[65107]: INFO nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Took 9.23 seconds to spawn the instance on the hypervisor. [ 691.864271] env[65107]: DEBUG nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 691.865663] env[65107]: DEBUG nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.871373] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd415202-5fa0-4ef7-bf06-e179963ffc48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.005987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.006324] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.006551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.006740] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.006918] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.010289] env[65107]: INFO nova.compute.manager [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Terminating instance [ 692.153042] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102440, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.377230] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.825s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.378182] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 692.382168] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.096s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.384365] env[65107]: INFO nova.compute.claims [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.400225] env[65107]: INFO nova.compute.manager [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Took 23.93 seconds to build instance. [ 692.521595] env[65107]: DEBUG nova.compute.manager [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 692.521595] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 692.523295] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efca008-0f93-4de1-9bc9-c388aa051c7e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.538020] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 692.539046] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35101d25-a370-4dca-868b-a2cf8c1868d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.551006] env[65107]: DEBUG oslo_vmware.api [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 692.551006] env[65107]: value = "task-5102441" [ 692.551006] env[65107]: _type = "Task" [ 692.551006] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.562875] env[65107]: DEBUG oslo_vmware.api [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.648656] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102440, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548238} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.649071] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96/71e6e280-50c3-4655-a6a3-40ea4c650d96.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 692.649319] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 692.649909] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57918976-090e-49ba-894c-9a79e535ec26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.659426] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 692.659426] env[65107]: value = "task-5102442" [ 692.659426] env[65107]: _type = "Task" [ 692.659426] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.676391] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.890154] env[65107]: DEBUG nova.compute.utils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 692.900032] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 692.900032] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 692.900032] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.900032] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.900032] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.900378] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.909810] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5dc21116-2ec8-4a41-8e2d-ba9b8bcde260 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.458s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.972047] env[65107]: DEBUG nova.policy [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4b2c7e86b0747ff8832c61ee96e1f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24d8b15ee5034b33a4c741412bcc71e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 693.066975] env[65107]: DEBUG oslo_vmware.api [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102441, 'name': PowerOffVM_Task, 'duration_secs': 0.396209} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.067350] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 693.067529] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 693.067824] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fe3b838-d94d-4734-bfc4-e806277d1613 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.141389] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 693.141706] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 693.141812] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Deleting the datastore file [datastore1] e40dc8ad-31cf-4798-ab97-3ee67c8fbe67 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 693.142109] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-023d02b9-36ce-484b-b2f6-bd2dd1641f6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.149679] env[65107]: DEBUG oslo_vmware.api [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for the task: (returnval){ [ 693.149679] env[65107]: value = "task-5102444" [ 693.149679] env[65107]: _type = "Task" [ 693.149679] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.160470] env[65107]: DEBUG oslo_vmware.api [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.169583] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113643} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.169844] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 693.170730] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5666d1dd-3fda-4ac5-94cf-ac65a5806db6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.196418] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96/71e6e280-50c3-4655-a6a3-40ea4c650d96.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 693.197242] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d566eab-798d-4f1c-82b5-805f4533b276 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.219089] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 693.219089] env[65107]: value = "task-5102445" [ 693.219089] env[65107]: _type = "Task" [ 693.219089] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.230655] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102445, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.398175] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 693.415035] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 693.518819] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Successfully created port: 232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 693.660759] env[65107]: DEBUG oslo_vmware.api [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Task: {'id': task-5102444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159522} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.663669] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 693.663948] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 693.664192] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.664495] env[65107]: INFO nova.compute.manager [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Took 1.14 seconds to destroy the instance on the hypervisor. [ 693.664743] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 693.665176] env[65107]: DEBUG nova.compute.manager [-] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 693.665302] env[65107]: DEBUG nova.network.neutron [-] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 693.665522] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.666143] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.666441] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.735622] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102445, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.947343] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.969588] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7c3c43-12f7-46b2-bc49-cc9656cd7c36 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.979167] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74876a1-78dc-475b-923f-d5d03313f190 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.014519] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07522882-8842-428a-8f58-27dfabb5475d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.024619] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8a7714-298c-448e-ba92-045744deec66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.041332] env[65107]: DEBUG nova.compute.provider_tree [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.242783] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102445, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.410209] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 694.566741] env[65107]: ERROR nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [req-3ec84aa2-9c7f-4ea1-a74d-373036baa81f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3ec84aa2-9c7f-4ea1-a74d-373036baa81f"}]} [ 694.590904] env[65107]: DEBUG nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 694.612551] env[65107]: DEBUG nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 694.612824] env[65107]: DEBUG nova.compute.provider_tree [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.627549] env[65107]: DEBUG nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 694.649032] env[65107]: DEBUG nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 694.737254] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102445, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.054845] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 695.150974] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdbb5b2-18b5-45c8-a932-5020f7776715 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.160936] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9255ccd-33ce-4c98-8075-45c1f41ebd71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.165721] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Successfully updated port: 232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 695.199908] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166a4e41-2dc7-4cf4-bfbe-b1ef9f3ead92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.209438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803e2aa1-6c49-4c43-b6c1-fb7343de31c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.225304] env[65107]: DEBUG nova.compute.provider_tree [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 695.237135] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102445, 'name': ReconfigVM_Task, 'duration_secs': 1.718242} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.237437] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96/71e6e280-50c3-4655-a6a3-40ea4c650d96.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 695.238735] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d0a7d00-a5a6-4f21-be34-9ec75d4f7e8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.247150] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 695.247150] env[65107]: value = "task-5102446" [ 695.247150] env[65107]: _type = "Task" [ 695.247150] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.260839] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102446, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.673242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "refresh_cache-07b83b81-9c65-4180-b208-c96525b78d8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.673242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "refresh_cache-07b83b81-9c65-4180-b208-c96525b78d8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.673242] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 695.761846] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102446, 'name': Rename_Task, 'duration_secs': 0.308878} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.761846] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 695.761846] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e7f46f3-a43b-45a5-a7ea-371324c56fc6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.772610] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 695.772610] env[65107]: value = "task-5102447" [ 695.772610] env[65107]: _type = "Task" [ 695.772610] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.786219] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.789108] env[65107]: DEBUG nova.scheduler.client.report [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 52 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 695.789378] env[65107]: DEBUG nova.compute.provider_tree [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 52 to 53 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 695.790183] env[65107]: DEBUG nova.compute.provider_tree [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 696.173059] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.173493] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.213370] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 696.234705] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.235298] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.291119] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102447, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.296985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.914s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.296985] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 696.299880] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.229s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.312778] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.313625] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.313913] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.430029] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Updating instance_info_cache with network_info: [{"id": "232c23c9-ffb2-48f4-8094-39a295742f87", "address": "fa:16:3e:60:d1:31", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap232c23c9-ff", "ovs_interfaceid": "232c23c9-ffb2-48f4-8094-39a295742f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.660359] env[65107]: DEBUG nova.network.neutron [-] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.749640] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 696.749929] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 696.750185] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 696.750402] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 696.750551] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 696.750698] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 696.750933] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.751198] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 696.751324] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 696.751414] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 696.751577] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 696.756231] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7899d6b3-8b70-49b9-9051-accb85552015 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.767688] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac46e0b-8172-4d42-b96d-2fc131cd0f5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.776080] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e10295-7adf-db1c-f6a4-ade829428807/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 696.776080] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81873f40-d63b-4aed-9a0e-a8b89a15bb88 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.793800] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e10295-7adf-db1c-f6a4-ade829428807/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 696.794077] env[65107]: ERROR oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e10295-7adf-db1c-f6a4-ade829428807/disk-0.vmdk due to incomplete transfer. [ 696.794352] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bdfbfb5c-6131-4603-b150-f2bde4486774 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.799353] env[65107]: DEBUG oslo_vmware.api [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102447, 'name': PowerOnVM_Task, 'duration_secs': 0.809123} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.801463] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 696.801463] env[65107]: INFO nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 10.23 seconds to spawn the instance on the hypervisor. [ 696.801463] env[65107]: DEBUG nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 696.801463] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10aabf36-d0a8-4108-9508-e9b5c15485e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.804776] env[65107]: DEBUG nova.compute.utils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 696.818325] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 696.818541] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 696.822050] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.822050] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.822050] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.822050] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.830038] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e10295-7adf-db1c-f6a4-ade829428807/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 696.830038] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Uploaded image 91f5fcd9-ad64-48e2-9af0-1f64e8a6cf13 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 696.832557] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 696.835369] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-187e9f39-643d-432f-982c-ceacf67e0cef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.849384] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 696.849384] env[65107]: value = "task-5102448" [ 696.849384] env[65107]: _type = "Task" [ 696.849384] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.859595] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102448, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.922739] env[65107]: DEBUG nova.policy [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4b2c7e86b0747ff8832c61ee96e1f0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24d8b15ee5034b33a4c741412bcc71e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 696.933267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "refresh_cache-07b83b81-9c65-4180-b208-c96525b78d8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.933692] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Instance network_info: |[{"id": "232c23c9-ffb2-48f4-8094-39a295742f87", "address": "fa:16:3e:60:d1:31", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap232c23c9-ff", "ovs_interfaceid": "232c23c9-ffb2-48f4-8094-39a295742f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 696.934621] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:d1:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c13fd8bc-e797-42fe-94ed-6370d3467a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '232c23c9-ffb2-48f4-8094-39a295742f87', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 696.946173] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 696.947312] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 696.947312] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1fec23a-a2d6-4c2f-b379-d1df9524d1bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.971707] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 696.971707] env[65107]: value = "task-5102449" [ 696.971707] env[65107]: _type = "Task" [ 696.971707] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.981438] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102449, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.166146] env[65107]: INFO nova.compute.manager [-] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Took 3.50 seconds to deallocate network for instance. [ 697.323024] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 697.358323] env[65107]: INFO nova.compute.manager [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 28.71 seconds to build instance. [ 697.361763] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e3510c31-d6be-4e9f-a0a2-a662123861e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.361918] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b779b5ad-7d93-47c2-b824-6d76246c00f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.362184] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 8306ff90-b8bd-4270-8133-96abe483156b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.362359] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ccd68268-d0fc-406e-896f-c61b114dd75d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 697.362521] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance be518c1d-edd7-40ff-b7cc-3310885b07b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.362643] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance f2814dd6-8be5-49b2-a030-f57e452c2e6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.362779] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.362903] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363028] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance f58c3088-c821-4eec-be0a-81221debc98e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363150] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 1253473b-d4bc-49ac-9724-9aa1e7a8f038 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363257] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e40dc8ad-31cf-4798-ab97-3ee67c8fbe67 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363360] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 705b91ea-ccd2-41ce-a14f-2ea5b70f80df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363485] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 7b24c530-56bf-4666-96f9-e2dc66728f1a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 697.363630] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 0e6c9f2d-4412-4d8d-8348-6552a34ab349 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363703] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 7062b119-01a4-45f4-bb52-212178382271 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.363829] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance f7446d76-45c2-4e8b-981d-d37c230cf125 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.364447] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ad53146e-210d-4321-b819-90795c1af260 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.364575] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance a2b72bac-6806-4b7b-b972-32aea52f7c82 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.364690] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b20a2c71-514d-488a-8edb-8fb0274a894a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.364860] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 71e6e280-50c3-4655-a6a3-40ea4c650d96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.364971] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 07b83b81-9c65-4180-b208-c96525b78d8a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.365323] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.369071] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Successfully created port: abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 697.379089] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102448, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.482504] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102449, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.676233] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.866041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c85d7b8-7249-4bf8-a167-5bf847e9b118 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.230s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.866041] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102448, 'name': Destroy_Task, 'duration_secs': 0.61695} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.874862] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Destroyed the VM [ 697.874862] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 697.874862] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-219e01e0-731f-434c-9aa7-a9d6845415ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.877785] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ef417f32-8c0e-4a62-9078-cb2bf95b85f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 697.881820] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 697.881820] env[65107]: value = "task-5102450" [ 697.881820] env[65107]: _type = "Task" [ 697.881820] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.893231] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102450, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.954522] env[65107]: DEBUG nova.compute.manager [req-c26152b1-2361-41f5-a8d7-417fb3f7027b req-47654199-36d1-4713-9ab6-6971a829bcbd service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Received event network-vif-plugged-232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 697.954724] env[65107]: DEBUG oslo_concurrency.lockutils [req-c26152b1-2361-41f5-a8d7-417fb3f7027b req-47654199-36d1-4713-9ab6-6971a829bcbd service nova] Acquiring lock "07b83b81-9c65-4180-b208-c96525b78d8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.955404] env[65107]: DEBUG oslo_concurrency.lockutils [req-c26152b1-2361-41f5-a8d7-417fb3f7027b req-47654199-36d1-4713-9ab6-6971a829bcbd service nova] Lock "07b83b81-9c65-4180-b208-c96525b78d8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.955587] env[65107]: DEBUG oslo_concurrency.lockutils [req-c26152b1-2361-41f5-a8d7-417fb3f7027b req-47654199-36d1-4713-9ab6-6971a829bcbd service nova] Lock "07b83b81-9c65-4180-b208-c96525b78d8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.955782] env[65107]: DEBUG nova.compute.manager [req-c26152b1-2361-41f5-a8d7-417fb3f7027b req-47654199-36d1-4713-9ab6-6971a829bcbd service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] No waiting events found dispatching network-vif-plugged-232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 697.956027] env[65107]: WARNING nova.compute.manager [req-c26152b1-2361-41f5-a8d7-417fb3f7027b req-47654199-36d1-4713-9ab6-6971a829bcbd service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Received unexpected event network-vif-plugged-232c23c9-ffb2-48f4-8094-39a295742f87 for instance with vm_state building and task_state spawning. [ 697.985015] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102449, 'name': CreateVM_Task, 'duration_secs': 0.579869} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.985221] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 697.985706] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.986077] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.986259] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.986816] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 697.987097] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea248c5a-f743-459a-a148-0881334aa35d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.997060] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 697.997060] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520017b6-0dde-eede-e5b6-453996853206" [ 697.997060] env[65107]: _type = "Task" [ 697.997060] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.006323] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520017b6-0dde-eede-e5b6-453996853206, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.060645] env[65107]: DEBUG nova.compute.manager [req-60ddef00-c4d6-425c-b7a3-bd22fb1a2bda req-a6f47638-1d2e-49f9-9cdc-48693c2eaa14 service nova] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Received event network-vif-deleted-2275d82f-e805-4c7b-8656-ff9475e7f524 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 698.114847] env[65107]: DEBUG nova.compute.manager [None req-5eed9ffd-4ee5-4677-928d-fc023ab98196 tempest-ServerDiagnosticsV248Test-880666234 tempest-ServerDiagnosticsV248Test-880666234-project-admin] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 698.115017] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ddcfcd-cad5-41a3-8de2-512b8f8f0b61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.124238] env[65107]: INFO nova.compute.manager [None req-5eed9ffd-4ee5-4677-928d-fc023ab98196 tempest-ServerDiagnosticsV248Test-880666234 tempest-ServerDiagnosticsV248Test-880666234-project-admin] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Retrieving diagnostics [ 698.125417] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e8a908-7726-41e5-8c8a-da52d4f3a844 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.332803] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 698.367387] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 698.367387] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 698.367663] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 698.367851] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 698.367999] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 698.368580] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 698.368580] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.368580] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 698.368735] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 698.368895] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 698.369114] env[65107]: DEBUG nova.virt.hardware [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 698.369649] env[65107]: DEBUG nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 698.373496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955300ea-fb65-4b9b-a867-b834dc3c1672 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.378917] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance c7739dfd-88d6-46ac-b3a0-e7dfe9b00332 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 698.394737] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ccacab-c3c7-4901-bdfc-7643585ffa59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.418477] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102450, 'name': RemoveSnapshot_Task} progress is 78%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.508220] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520017b6-0dde-eede-e5b6-453996853206, 'name': SearchDatastore_Task, 'duration_secs': 0.014366} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.509247] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.509247] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.509247] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.509247] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.509451] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.509629] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb0928ce-5c04-4ad4-ad62-96930fd361ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.521476] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.521681] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.522554] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25dc1218-70dd-4cab-83d2-cceffeafafdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.530198] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 698.530198] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ffd66c-bffa-9708-3a89-a49247a02085" [ 698.530198] env[65107]: _type = "Task" [ 698.530198] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.541327] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ffd66c-bffa-9708-3a89-a49247a02085, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.889622] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3948dcdd-e7c1-40b6-a3ec-97741940a0fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 698.905933] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102450, 'name': RemoveSnapshot_Task} progress is 78%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.909333] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.042430] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ffd66c-bffa-9708-3a89-a49247a02085, 'name': SearchDatastore_Task, 'duration_secs': 0.011751} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.044524] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3839ab2-3167-4b55-8734-78ea49b646d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.051651] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 699.051651] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52951bc4-f63f-3bd9-8f89-5fc1e661229d" [ 699.051651] env[65107]: _type = "Task" [ 699.051651] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.061636] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52951bc4-f63f-3bd9-8f89-5fc1e661229d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.064861] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Successfully updated port: abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 699.264755] env[65107]: INFO nova.compute.manager [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Rescuing [ 699.266089] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.266520] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.268603] env[65107]: DEBUG nova.network.neutron [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 699.401363] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 699.403360] env[65107]: DEBUG oslo_vmware.api [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102450, 'name': RemoveSnapshot_Task, 'duration_secs': 1.106842} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.403723] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 699.405083] env[65107]: INFO nova.compute.manager [None req-5c42e509-e9a4-4d3f-bf23-e9ffb36006c0 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Took 16.61 seconds to snapshot the instance on the hypervisor. [ 699.569677] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "refresh_cache-23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.569677] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "refresh_cache-23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.569677] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 699.569677] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52951bc4-f63f-3bd9-8f89-5fc1e661229d, 'name': SearchDatastore_Task, 'duration_secs': 0.02586} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.569817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.569817] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 07b83b81-9c65-4180-b208-c96525b78d8a/07b83b81-9c65-4180-b208-c96525b78d8a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.573028] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79f7cb46-6ded-400b-ad9d-002f9cc4b93d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.580043] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 699.580043] env[65107]: value = "task-5102451" [ 699.580043] env[65107]: _type = "Task" [ 699.580043] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.594023] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.770405] env[65107]: WARNING neutronclient.v2_0.client [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 699.771204] env[65107]: WARNING openstack [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.771965] env[65107]: WARNING openstack [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.905798] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 90706f49-af87-4038-ae76-02a95e4a12e6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 699.996141] env[65107]: WARNING openstack [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.996879] env[65107]: WARNING openstack [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.071894] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.072415] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.081288] env[65107]: WARNING neutronclient.v2_0.client [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.081956] env[65107]: WARNING openstack [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.082317] env[65107]: WARNING openstack [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.104807] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102451, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.114679] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 700.154519] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.155166] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.222586] env[65107]: DEBUG nova.network.neutron [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 700.244983] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4c4ec-76d7-7d3d-addb-65be64c7f8e4/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 700.247841] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05226d5f-0076-41a8-ad81-f31d7e9a1de5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.260723] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4c4ec-76d7-7d3d-addb-65be64c7f8e4/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 700.260963] env[65107]: ERROR oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4c4ec-76d7-7d3d-addb-65be64c7f8e4/disk-0.vmdk due to incomplete transfer. [ 700.262568] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.263396] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.263828] env[65107]: WARNING openstack [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.273401] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-91f71442-0ef1-419e-a254-ba170fa8ad83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.287098] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4c4ec-76d7-7d3d-addb-65be64c7f8e4/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 700.287583] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Uploaded image c57c1202-2e52-423c-9825-ad7cbbb94e56 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 700.289929] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 700.290120] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f72980ea-8bca-4c52-868e-e1a2d5a00a17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.297834] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 700.297834] env[65107]: value = "task-5102452" [ 700.297834] env[65107]: _type = "Task" [ 700.297834] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.311685] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102452, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.373375] env[65107]: DEBUG nova.network.neutron [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Updating instance_info_cache with network_info: [{"id": "abb39b9b-672f-498e-b98a-38ab04a19143", "address": "fa:16:3e:d0:0d:ee", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabb39b9b-67", "ovs_interfaceid": "abb39b9b-672f-498e-b98a-38ab04a19143", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 700.409253] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 854f7a3f-9b7a-4506-8310-94533f0e4cfc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 700.464155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "a2b72bac-6806-4b7b-b972-32aea52f7c82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.464437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "a2b72bac-6806-4b7b-b972-32aea52f7c82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.465208] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "a2b72bac-6806-4b7b-b972-32aea52f7c82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.465415] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "a2b72bac-6806-4b7b-b972-32aea52f7c82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.465590] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "a2b72bac-6806-4b7b-b972-32aea52f7c82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 700.467615] env[65107]: INFO nova.compute.manager [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Terminating instance [ 700.602021] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.797762} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.602481] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 07b83b81-9c65-4180-b208-c96525b78d8a/07b83b81-9c65-4180-b208-c96525b78d8a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.602639] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.602907] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8c95abb-64cf-4ab0-8683-18ad9dae4a67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.611300] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 700.611300] env[65107]: value = "task-5102453" [ 700.611300] env[65107]: _type = "Task" [ 700.611300] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.623315] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.726920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.812756] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102452, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.878038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "refresh_cache-23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.878038] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Instance network_info: |[{"id": "abb39b9b-672f-498e-b98a-38ab04a19143", "address": "fa:16:3e:d0:0d:ee", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabb39b9b-67", "ovs_interfaceid": "abb39b9b-672f-498e-b98a-38ab04a19143", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 700.878273] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:0d:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c13fd8bc-e797-42fe-94ed-6370d3467a7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'abb39b9b-672f-498e-b98a-38ab04a19143', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.886999] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 700.887684] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 700.887960] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31d527c9-64ae-4bf2-95f0-80c55b3157ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.912729] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d40aec5a-d898-42f3-a121-503cf043f72b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 700.918045] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.918045] env[65107]: value = "task-5102454" [ 700.918045] env[65107]: _type = "Task" [ 700.918045] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.927453] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102454, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.972666] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "refresh_cache-a2b72bac-6806-4b7b-b972-32aea52f7c82" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.972666] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquired lock "refresh_cache-a2b72bac-6806-4b7b-b972-32aea52f7c82" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.972815] env[65107]: DEBUG nova.network.neutron [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 701.024383] env[65107]: DEBUG nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Received event network-changed-232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 701.024609] env[65107]: DEBUG nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Refreshing instance network info cache due to event network-changed-232c23c9-ffb2-48f4-8094-39a295742f87. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 701.024829] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Acquiring lock "refresh_cache-07b83b81-9c65-4180-b208-c96525b78d8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.024962] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Acquired lock "refresh_cache-07b83b81-9c65-4180-b208-c96525b78d8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.025154] env[65107]: DEBUG nova.network.neutron [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Refreshing network info cache for port 232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 701.123928] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074724} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.124136] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.125329] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7213fe22-faf6-47cd-85b0-3fa4ba7a10fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.152738] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] 07b83b81-9c65-4180-b208-c96525b78d8a/07b83b81-9c65-4180-b208-c96525b78d8a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.153141] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4be946d2-02a4-4fe5-959b-3580f0e54c53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.179528] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 701.179528] env[65107]: value = "task-5102455" [ 701.179528] env[65107]: _type = "Task" [ 701.179528] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.193293] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102455, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.314981] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102452, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.418036] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 701.430871] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102454, 'name': CreateVM_Task, 'duration_secs': 0.46182} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.431782] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 701.432366] env[65107]: WARNING neutronclient.v2_0.client [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.432859] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.433018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.433297] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 701.434951] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef652b01-bdcb-421a-8e6b-d1c6801e2721 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.439380] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 701.439380] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8f70c-c919-a791-9c2b-86a026b1a7c3" [ 701.439380] env[65107]: _type = "Task" [ 701.439380] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.450037] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.450037] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.456133] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8f70c-c919-a791-9c2b-86a026b1a7c3, 'name': SearchDatastore_Task, 'duration_secs': 0.011275} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.456491] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.456716] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.457386] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.457386] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.457386] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.457549] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f206a6a3-4866-4cb1-b6cb-af81ef833b86 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.467577] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.467827] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 701.468492] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4634c14-850e-44eb-b279-65d8effa9a9d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.475557] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 701.475557] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8192-1034-d97a-7b6b-8e9103448e8c" [ 701.475557] env[65107]: _type = "Task" [ 701.475557] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.478737] env[65107]: WARNING neutronclient.v2_0.client [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.479743] env[65107]: WARNING openstack [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.480171] env[65107]: WARNING openstack [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.495684] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8192-1034-d97a-7b6b-8e9103448e8c, 'name': SearchDatastore_Task, 'duration_secs': 0.010603} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.496781] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67798d62-fd87-4608-9961-4938be56b1b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.505313] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 701.505313] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5266d33a-89f4-7b59-bd8e-9062b23a86a7" [ 701.505313] env[65107]: _type = "Task" [ 701.505313] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.510661] env[65107]: DEBUG nova.network.neutron [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 701.515833] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5266d33a-89f4-7b59-bd8e-9062b23a86a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.532027] env[65107]: WARNING neutronclient.v2_0.client [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.532027] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.532027] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.615762] env[65107]: DEBUG nova.network.neutron [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 701.650637] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "bfe544ec-9b08-4118-a940-a51520ecaac0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.650883] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.660089] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.660493] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.691798] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102455, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.727266] env[65107]: WARNING neutronclient.v2_0.client [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.727944] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.728310] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.737856] env[65107]: DEBUG nova.compute.manager [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Received event network-changed-3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 701.738052] env[65107]: DEBUG nova.compute.manager [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Refreshing instance network info cache due to event network-changed-3d38d576-1871-49a5-bc10-aef1a37abfa0. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 701.738260] env[65107]: DEBUG oslo_concurrency.lockutils [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.738396] env[65107]: DEBUG oslo_concurrency.lockutils [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.738548] env[65107]: DEBUG nova.network.neutron [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Refreshing network info cache for port 3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 701.812180] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102452, 'name': Destroy_Task, 'duration_secs': 1.081925} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.812544] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Destroyed the VM [ 701.812820] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 701.813133] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fae45372-3a73-45da-8bbb-da5f169479ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.818438] env[65107]: DEBUG nova.network.neutron [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Updated VIF entry in instance network info cache for port 232c23c9-ffb2-48f4-8094-39a295742f87. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 701.818920] env[65107]: DEBUG nova.network.neutron [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Updating instance_info_cache with network_info: [{"id": "232c23c9-ffb2-48f4-8094-39a295742f87", "address": "fa:16:3e:60:d1:31", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap232c23c9-ff", "ovs_interfaceid": "232c23c9-ffb2-48f4-8094-39a295742f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 701.821636] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 701.821636] env[65107]: value = "task-5102456" [ 701.821636] env[65107]: _type = "Task" [ 701.821636] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.831885] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102456, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.921653] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 369caf24-48a6-4640-8601-e30fd47453da has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 702.015909] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5266d33a-89f4-7b59-bd8e-9062b23a86a7, 'name': SearchDatastore_Task, 'duration_secs': 0.011106} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.016518] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.016650] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6/23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 702.017025] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-947c9a9c-88a2-40f6-86c5-1c9ff47012d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.025203] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 702.025203] env[65107]: value = "task-5102457" [ 702.025203] env[65107]: _type = "Task" [ 702.025203] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.035011] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.118483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Releasing lock "refresh_cache-a2b72bac-6806-4b7b-b972-32aea52f7c82" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.119383] env[65107]: DEBUG nova.compute.manager [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 702.119636] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.120681] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44bbf67-a881-4607-a757-610b5f449fe8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.129588] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 702.129831] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-186d10af-dd70-432c-9b89-9d201e9abf11 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.138203] env[65107]: DEBUG oslo_vmware.api [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 702.138203] env[65107]: value = "task-5102458" [ 702.138203] env[65107]: _type = "Task" [ 702.138203] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.148340] env[65107]: DEBUG oslo_vmware.api [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102458, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.192517] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102455, 'name': ReconfigVM_Task, 'duration_secs': 0.811893} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.194075] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Reconfigured VM instance instance-0000001d to attach disk [datastore1] 07b83b81-9c65-4180-b208-c96525b78d8a/07b83b81-9c65-4180-b208-c96525b78d8a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.194075] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14329596-23e6-49b3-a1c1-b090b531fc3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.202658] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 702.202658] env[65107]: value = "task-5102459" [ 702.202658] env[65107]: _type = "Task" [ 702.202658] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.213217] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102459, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.243114] env[65107]: WARNING neutronclient.v2_0.client [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 702.243939] env[65107]: WARNING openstack [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.244361] env[65107]: WARNING openstack [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.268942] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 702.269181] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b9238fa-b029-40ce-8dd4-a3441b6c897b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.279624] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 702.279624] env[65107]: value = "task-5102460" [ 702.279624] env[65107]: _type = "Task" [ 702.279624] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.294056] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102460, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.329540] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Releasing lock "refresh_cache-07b83b81-9c65-4180-b208-c96525b78d8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.329976] env[65107]: DEBUG nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Received event network-vif-plugged-abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 702.330317] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Acquiring lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.330689] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.330944] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.331220] env[65107]: DEBUG nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] No waiting events found dispatching network-vif-plugged-abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 702.331553] env[65107]: WARNING nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Received unexpected event network-vif-plugged-abb39b9b-672f-498e-b98a-38ab04a19143 for instance with vm_state building and task_state spawning. [ 702.332018] env[65107]: DEBUG nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Received event network-changed-abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 702.332250] env[65107]: DEBUG nova.compute.manager [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Refreshing instance network info cache due to event network-changed-abb39b9b-672f-498e-b98a-38ab04a19143. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 702.332488] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Acquiring lock "refresh_cache-23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.332763] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Acquired lock "refresh_cache-23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.333036] env[65107]: DEBUG nova.network.neutron [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Refreshing network info cache for port abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 702.347647] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102456, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.406431] env[65107]: WARNING openstack [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.407712] env[65107]: WARNING openstack [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.425341] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 702.426780] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 702.426780] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4416MB phys_disk=100GB used_disk=19GB total_vcpus=48 used_vcpus=20 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '20', 'num_vm_error': '1', 'num_task_None': '15', 'num_os_type_None': '20', 'num_proj_a2d4844feda0432ebe1ef160b0070b83': '4', 'io_workload': '3', 'num_vm_active': '16', 'num_proj_2882c2c7949045519a146655bb694a3f': '1', 'num_proj_d98fae2faa0a40c798e251737ccb31c6': '1', 'num_proj_634f89d91e1443c7a0eb1314ecb17232': '1', 'num_proj_8492fd3bcbf0485c84e4bf980f3d2809': '2', 'num_task_image_uploading': '2', 'num_proj_bcb6424259344f3fa87ba9a4c10637d2': '1', 'num_task_deleting': '1', 'num_proj_520207ef29e64aa799841fae36e469f5': '1', 'num_proj_9c5372cbba21404eb54eda5f08f0ac42': '1', 'num_proj_e9fcd71f4f064b488f832406fedafbe0': '1', 'num_proj_e3fa2b26480f4495926fadf02489de3e': '1', 'num_proj_95393e0b56bd47cd8bb37207e0500269': '1', 'num_proj_ba25b2abe65f41a98d8db49cbc2a1e2d': '1', 'num_proj_a4457697290e40e89d4261fe02ff5eb7': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_f379144b78764fe394039d87b043a946': '1', 'num_proj_24d8b15ee5034b33a4c741412bcc71e0': '2'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 702.499578] env[65107]: WARNING neutronclient.v2_0.client [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 702.500510] env[65107]: WARNING openstack [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.500724] env[65107]: WARNING openstack [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.538503] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102457, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.607174] env[65107]: DEBUG nova.network.neutron [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updated VIF entry in instance network info cache for port 3d38d576-1871-49a5-bc10-aef1a37abfa0. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 702.607699] env[65107]: DEBUG nova.network.neutron [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 702.654649] env[65107]: DEBUG oslo_vmware.api [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102458, 'name': PowerOffVM_Task, 'duration_secs': 0.164437} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.655107] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 702.655107] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 702.655858] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-870b5b6e-e773-4637-b2b2-2160f92d4a77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.681701] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 702.682052] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 702.682146] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Deleting the datastore file [datastore1] a2b72bac-6806-4b7b-b972-32aea52f7c82 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 702.682400] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9cfc3fa-dbfd-4af5-8d4a-c75afedf44da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.692369] env[65107]: DEBUG oslo_vmware.api [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for the task: (returnval){ [ 702.692369] env[65107]: value = "task-5102462" [ 702.692369] env[65107]: _type = "Task" [ 702.692369] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.699896] env[65107]: DEBUG oslo_vmware.api [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.713202] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102459, 'name': Rename_Task, 'duration_secs': 0.213625} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.713474] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.713716] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc54eeb1-7171-4605-9af8-98bf4e3a20ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.723613] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 702.723613] env[65107]: value = "task-5102463" [ 702.723613] env[65107]: _type = "Task" [ 702.723613] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.733208] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.790620] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102460, 'name': PowerOffVM_Task, 'duration_secs': 0.229474} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.794263] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 702.795736] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11c0717-09b3-49c5-b94d-116a077fda31 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.819871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6bd3d0-95fe-40bb-afde-d31c62605fb4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.832303] env[65107]: DEBUG oslo_vmware.api [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102456, 'name': RemoveSnapshot_Task, 'duration_secs': 0.703006} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.835031] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 702.835349] env[65107]: INFO nova.compute.manager [None req-5eb3ba97-8a98-4d16-9525-27784189e678 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Took 17.62 seconds to snapshot the instance on the hypervisor. [ 702.845486] env[65107]: WARNING neutronclient.v2_0.client [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 702.847630] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.850027] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.881200] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 702.881618] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c44f8d5-e69a-4ebe-9483-77f9c15c5214 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.895474] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 702.895474] env[65107]: value = "task-5102464" [ 702.895474] env[65107]: _type = "Task" [ 702.895474] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.908610] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102464, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.999120] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.999520] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.043931] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.633366} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.046607] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6/23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 703.046898] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 703.047860] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1934399f-ea48-42fc-b528-cf71ad0fe214 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.055148] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 703.055148] env[65107]: value = "task-5102465" [ 703.055148] env[65107]: _type = "Task" [ 703.055148] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.068136] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.073120] env[65107]: WARNING neutronclient.v2_0.client [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.073744] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.074126] env[65107]: WARNING openstack [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.083796] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a61bfa0-d8eb-4ab8-9631-f58137819dc9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.092205] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5d6035-eb0a-48e7-bf8f-a97d854a22a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.128256] env[65107]: DEBUG oslo_concurrency.lockutils [req-66dbf44c-8881-44a8-a1a0-1648ec4a6cef req-14427d42-fc64-4c93-91ea-776138f8cf03 service nova] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.135117] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d6426e-024c-4c64-a561-7fb5bafdb277 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.144382] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc957ea5-35a7-4236-bff9-9f2f18537569 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.162289] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.186080] env[65107]: DEBUG nova.network.neutron [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Updated VIF entry in instance network info cache for port abb39b9b-672f-498e-b98a-38ab04a19143. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 703.186080] env[65107]: DEBUG nova.network.neutron [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Updating instance_info_cache with network_info: [{"id": "abb39b9b-672f-498e-b98a-38ab04a19143", "address": "fa:16:3e:d0:0d:ee", "network": {"id": "f07c9962-99ac-4463-b5b8-0066c539c3a9", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-2142986147-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24d8b15ee5034b33a4c741412bcc71e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c13fd8bc-e797-42fe-94ed-6370d3467a7f", "external-id": "nsx-vlan-transportzone-833", "segmentation_id": 833, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabb39b9b-67", "ovs_interfaceid": "abb39b9b-672f-498e-b98a-38ab04a19143", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.201485] env[65107]: DEBUG oslo_vmware.api [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Task: {'id': task-5102462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206961} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.201895] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 703.201971] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 703.202133] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.202306] env[65107]: INFO nova.compute.manager [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Took 1.08 seconds to destroy the instance on the hypervisor. [ 703.202553] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 703.202732] env[65107]: DEBUG nova.compute.manager [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 703.202828] env[65107]: DEBUG nova.network.neutron [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 703.203091] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.203603] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.203856] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.230429] env[65107]: DEBUG nova.network.neutron [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 703.230704] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.240968] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102463, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.410026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 703.411131] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 703.411131] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.411296] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.411605] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 703.412035] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2d9c2c1-76b7-4b0e-a15c-459d3d2b5382 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.426032] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 703.428041] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 703.428041] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6517fcf-c607-431c-b297-a4d35cf2b6b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.436391] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 703.436391] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52089728-ee7a-59e6-6312-637598786c42" [ 703.436391] env[65107]: _type = "Task" [ 703.436391] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.445352] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52089728-ee7a-59e6-6312-637598786c42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.564774] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.274625} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.565088] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.565885] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58247e0-61e7-483f-a9d9-22deb12cc9e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.589142] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6/23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.589659] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e6d3e21-2f82-4b4c-a9fa-adf175692087 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.610787] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 703.610787] env[65107]: value = "task-5102466" [ 703.610787] env[65107]: _type = "Task" [ 703.610787] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.619511] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102466, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.688026] env[65107]: ERROR nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [req-ab1a3b3a-77f9-49d0-9672-b27385dd5da3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ab1a3b3a-77f9-49d0-9672-b27385dd5da3"}]} [ 703.690220] env[65107]: DEBUG oslo_concurrency.lockutils [req-b4afeeaa-9161-4f33-8e5a-39deb6bf02d5 req-7c7fb323-d88c-42e4-a7bb-39a5901ba401 service nova] Releasing lock "refresh_cache-23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.707796] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 703.726996] env[65107]: INFO nova.compute.manager [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Rebuilding instance [ 703.734517] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 703.734781] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.737043] env[65107]: DEBUG nova.network.neutron [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.745599] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102463, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.750227] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 703.775410] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 703.786144] env[65107]: DEBUG nova.compute.manager [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 703.787040] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b711314-1dcb-4c47-b3fb-5d6a9ccc03dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.953483] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52089728-ee7a-59e6-6312-637598786c42, 'name': SearchDatastore_Task, 'duration_secs': 0.015949} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.954894] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e54b028-f005-46e8-9bbc-e68911e90df3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.967653] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 703.967653] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529e6dc6-1cb1-c90e-1051-647c24d7a98d" [ 703.967653] env[65107]: _type = "Task" [ 703.967653] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.977254] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529e6dc6-1cb1-c90e-1051-647c24d7a98d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.121278] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102466, 'name': ReconfigVM_Task, 'duration_secs': 0.323839} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.124620] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6/23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.124881] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c6af869-e797-44cd-ab3a-e4b5b9b46c24 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.132020] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 704.132020] env[65107]: value = "task-5102467" [ 704.132020] env[65107]: _type = "Task" [ 704.132020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.142877] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102467, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.237548] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102463, 'name': PowerOnVM_Task, 'duration_secs': 1.242742} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.240608] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 704.241133] env[65107]: INFO nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Took 9.83 seconds to spawn the instance on the hypervisor. [ 704.241435] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 704.241855] env[65107]: INFO nova.compute.manager [-] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Took 1.04 seconds to deallocate network for instance. [ 704.242939] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1644178-c1ef-4881-86fe-f2999bb33d03 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.349845] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2afd57a-c8fc-412e-9433-b6ce3f07bd42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.358830] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0fcb74-5284-4869-8452-fdc7e10425ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.394902] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f218be27-3417-41f0-b37c-e00375d22519 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.407263] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01add061-7c13-4f72-a12f-0360d75b89d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.421948] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 704.478471] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529e6dc6-1cb1-c90e-1051-647c24d7a98d, 'name': SearchDatastore_Task, 'duration_secs': 0.011339} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.478761] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.479026] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. {{(pid=65107) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 704.479296] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-119741db-60fa-4a79-ab82-c544cf497944 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.487568] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 704.487568] env[65107]: value = "task-5102468" [ 704.487568] env[65107]: _type = "Task" [ 704.487568] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.498423] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.645501] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102467, 'name': Rename_Task, 'duration_secs': 0.152056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.646105] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 704.646237] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81586f42-b72e-4ffd-8214-7176f477ad84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.659496] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 704.659496] env[65107]: value = "task-5102469" [ 704.659496] env[65107]: _type = "Task" [ 704.659496] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.671496] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.760479] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.770192] env[65107]: INFO nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Took 32.53 seconds to build instance. [ 704.806536] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 704.807019] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef4d5552-5a49-4949-a67a-eeec266af330 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.816487] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 704.816487] env[65107]: value = "task-5102470" [ 704.816487] env[65107]: _type = "Task" [ 704.816487] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.831735] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.959403] env[65107]: DEBUG nova.compute.manager [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 704.962784] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e2d692-a30b-4876-8d35-892fb7786d04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.990254] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 55 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 704.990575] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 55 to 56 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 704.990763] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 705.007393] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102468, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.173293] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102469, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.272768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "07b83b81-9c65-4180-b208-c96525b78d8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.063s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.334029] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102470, 'name': PowerOffVM_Task, 'duration_secs': 0.357966} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.334029] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 705.334029] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.334029] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7242fef0-7620-48d8-a112-d7dd63aafb39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.343109] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 705.343109] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f6e3276-47cf-4d54-9b41-07f6940ca659 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.414860] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 705.415098] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 705.415292] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleting the datastore file [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 705.416380] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de15eeb7-cb3c-49ee-be9c-7970e33fde07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.423080] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 705.423080] env[65107]: value = "task-5102472" [ 705.423080] env[65107]: _type = "Task" [ 705.423080] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.435397] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.476788] env[65107]: INFO nova.compute.manager [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] instance snapshotting [ 705.479741] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad0a2f2-334f-412e-849d-70fbf7b89b9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.501800] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 705.502100] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.202s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.505643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.592s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.507187] env[65107]: INFO nova.compute.claims [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.510501] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d81d6f-69c9-4678-a7da-dd4e65880019 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.519864] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545448} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.522019] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. [ 705.526043] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8abb4a-8a68-4bf7-a8fd-83d9947d0b71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.557233] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 705.557600] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be7faf3a-5562-49cd-bb4b-f6dfc2a67f96 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.580661] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 705.580661] env[65107]: value = "task-5102473" [ 705.580661] env[65107]: _type = "Task" [ 705.580661] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.590438] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102473, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.669491] env[65107]: DEBUG oslo_vmware.api [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102469, 'name': PowerOnVM_Task, 'duration_secs': 0.582382} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.669802] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 705.670057] env[65107]: INFO nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Took 7.34 seconds to spawn the instance on the hypervisor. [ 705.670246] env[65107]: DEBUG nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 705.671051] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597a77a4-5da5-4edb-8e9d-afbb9f9e1491 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.775825] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 705.925206] env[65107]: DEBUG nova.compute.manager [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 705.928029] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2b3f2a-52f5-408a-8f9d-13b42d5a5c10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.939734] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15793} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.941780] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 705.941979] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 705.942244] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.030588] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 706.031358] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2261dc59-a7c6-4b94-961d-4c535b6ab534 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.045142] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 706.045142] env[65107]: value = "task-5102474" [ 706.045142] env[65107]: _type = "Task" [ 706.045142] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.057205] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102474, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.096069] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102473, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.195275] env[65107]: INFO nova.compute.manager [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Took 33.93 seconds to build instance. [ 706.300982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.452197] env[65107]: INFO nova.compute.manager [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] instance snapshotting [ 706.455479] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca38911-8436-43e3-8945-486cef372207 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.475894] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b937b3d-e5e8-4bfa-a49d-bb2cee421abe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.556124] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102474, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.593792] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102473, 'name': ReconfigVM_Task, 'duration_secs': 0.934521} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.596695] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Reconfigured VM instance instance-0000001b to attach disk [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 706.597956] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473b4286-5c79-4ade-a06a-84a8785164cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.628302] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92f607bc-3395-45c7-b8be-186ab813d767 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.645252] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 706.645252] env[65107]: value = "task-5102475" [ 706.645252] env[65107]: _type = "Task" [ 706.645252] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.656948] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102475, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.695577] env[65107]: DEBUG oslo_concurrency.lockutils [None req-836e1bff-0c30-48eb-8266-2cdedebb9ea8 tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.449s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.989196] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 706.992022] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 706.992295] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 706.992453] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 706.992630] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 706.992773] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 706.992936] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 706.993163] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.993321] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 706.993486] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 706.993645] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 706.993821] env[65107]: DEBUG nova.virt.hardware [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 706.994249] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0c9cc71c-bbd3-405b-95e7-de9addc56fa3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.999210] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f45a862-5234-4751-8f7a-5eb8f1841598 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.012273] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4117fa32-056a-4a14-8a62-53e08520f8f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.016858] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 707.016858] env[65107]: value = "task-5102476" [ 707.016858] env[65107]: _type = "Task" [ 707.016858] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.031526] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:7b:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2686e3e0-c528-42bb-9444-050209faa6b1', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 707.039780] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 707.046478] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 707.046842] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102476, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.047362] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d1a7c18-2c30-4829-aca0-235a5cf9e0e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.078345] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102474, 'name': CreateSnapshot_Task, 'duration_secs': 1.00594} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.079910] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 707.081386] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 707.081386] env[65107]: value = "task-5102477" [ 707.081386] env[65107]: _type = "Task" [ 707.081386] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.081386] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cfb257-ba37-4bfc-bce8-b0daefef6c29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.095360] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102477, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.157248] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102475, 'name': ReconfigVM_Task, 'duration_secs': 0.224429} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.160045] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.160707] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78f3e0cf-45a3-443f-b637-b410f97b8881 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.168827] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 707.168827] env[65107]: value = "task-5102478" [ 707.168827] env[65107]: _type = "Task" [ 707.168827] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.174439] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365832a4-b489-4530-8c3e-5ad1aadd4cdc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.187070] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd80ad2-d431-4adf-8370-1d0f9d5ae566 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.190976] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102478, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.222825] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 707.226825] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1a2831-0aea-463d-805f-ecb583580847 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.236176] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937801fd-78c9-4885-937f-18ca55379198 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.252090] env[65107]: DEBUG nova.compute.provider_tree [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.533821] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102476, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.597231] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102477, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.614100] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 707.614458] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-96aede70-1ce2-4ee1-a96b-e6b8828937db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.623796] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 707.623796] env[65107]: value = "task-5102479" [ 707.623796] env[65107]: _type = "Task" [ 707.623796] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.633926] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102479, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.682285] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102478, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.751806] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.755133] env[65107]: DEBUG nova.scheduler.client.report [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 708.032546] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102476, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.096058] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102477, 'name': CreateVM_Task, 'duration_secs': 0.615986} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.096380] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 708.097033] env[65107]: WARNING neutronclient.v2_0.client [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 708.097482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.097643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 708.097977] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 708.098295] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a678385-9e53-4990-a955-2092e17d8e1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.104532] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 708.104532] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52544830-5a1b-d213-bac8-7da25cb47bb0" [ 708.104532] env[65107]: _type = "Task" [ 708.104532] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.117934] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52544830-5a1b-d213-bac8-7da25cb47bb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.137128] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102479, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.184885] env[65107]: DEBUG oslo_vmware.api [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102478, 'name': PowerOnVM_Task, 'duration_secs': 0.881813} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.185206] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.190791] env[65107]: DEBUG nova.compute.manager [None req-c0083909-2448-430e-b64f-d76173abc7a3 tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 708.191633] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a06b458-4ac3-4b84-89e1-4cadf989b4ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.261180] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.755s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 708.261451] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 708.264212] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.491s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.264430] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 708.269961] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.385s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.271808] env[65107]: INFO nova.compute.claims [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.306961] env[65107]: INFO nova.scheduler.client.report [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Deleted allocations for instance 7b24c530-56bf-4666-96f9-e2dc66728f1a [ 708.530595] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102476, 'name': CreateSnapshot_Task, 'duration_secs': 1.028058} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.531049] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 708.533918] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6411a0-a7f1-45ed-9b9d-525a6670f598 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.615752] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52544830-5a1b-d213-bac8-7da25cb47bb0, 'name': SearchDatastore_Task, 'duration_secs': 0.013293} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.616114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.616331] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.616647] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.620026] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 708.620026] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.620026] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27ef6d51-612d-4672-b89d-4e69a70e4835 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.631648] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.631900] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 708.633678] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71786d62-88fa-400f-9a10-bb7ff0dd9328 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.641495] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102479, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.645918] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 708.645918] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f6a938-e77b-8dde-5e1d-5bfbc757f773" [ 708.645918] env[65107]: _type = "Task" [ 708.645918] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.656542] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f6a938-e77b-8dde-5e1d-5bfbc757f773, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.747954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.748802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.748802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.748802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.748987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 708.751188] env[65107]: INFO nova.compute.manager [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Terminating instance [ 708.776481] env[65107]: DEBUG nova.compute.utils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 708.780683] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 708.781142] env[65107]: DEBUG nova.network.neutron [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 708.781512] env[65107]: WARNING neutronclient.v2_0.client [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.782421] env[65107]: WARNING neutronclient.v2_0.client [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.783760] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.784039] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.815254] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a4ea1cbe-7618-4857-96b1-b0b08b5c4947 tempest-DeleteServersAdminTestJSON-453173968 tempest-DeleteServersAdminTestJSON-453173968-project-member] Lock "7b24c530-56bf-4666-96f9-e2dc66728f1a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.669s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 708.843899] env[65107]: DEBUG nova.policy [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75dbe2746e84cc2ad6848d6bd5412b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58588fe2a11f4ee0aab313903b211fa8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 709.062232] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 709.065379] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-85364702-d783-4a3a-9560-dfebe3559bce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.074796] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 709.074796] env[65107]: value = "task-5102480" [ 709.074796] env[65107]: _type = "Task" [ 709.074796] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.095218] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102480, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.140115] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102479, 'name': CloneVM_Task, 'duration_secs': 1.501415} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.143456] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Created linked-clone VM from snapshot [ 709.144743] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b15d77-f5c3-4496-a640-d82d52972421 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.161030] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Uploading image 39f72a8a-740a-4530-923e-597c889d6992 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 709.170493] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f6a938-e77b-8dde-5e1d-5bfbc757f773, 'name': SearchDatastore_Task, 'duration_secs': 0.012844} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.172320] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8c34a79-b221-464c-adbf-04595ed35eac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.179804] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 709.179804] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f02f74-d53a-c234-c3d7-e446a81f4f85" [ 709.179804] env[65107]: _type = "Task" [ 709.179804] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.193769] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f02f74-d53a-c234-c3d7-e446a81f4f85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.206479] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 709.206479] env[65107]: value = "vm-992666" [ 709.206479] env[65107]: _type = "VirtualMachine" [ 709.206479] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 709.207755] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ade8222a-ccd8-4070-9b2d-b66e046dfa77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.220221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "07b83b81-9c65-4180-b208-c96525b78d8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.220636] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "07b83b81-9c65-4180-b208-c96525b78d8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.220890] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "07b83b81-9c65-4180-b208-c96525b78d8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.221108] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "07b83b81-9c65-4180-b208-c96525b78d8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.221367] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "07b83b81-9c65-4180-b208-c96525b78d8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.224312] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lease: (returnval){ [ 709.224312] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525599ab-4a17-623a-7256-07ab9aa4f959" [ 709.224312] env[65107]: _type = "HttpNfcLease" [ 709.224312] env[65107]: } obtained for exporting VM: (result){ [ 709.224312] env[65107]: value = "vm-992666" [ 709.224312] env[65107]: _type = "VirtualMachine" [ 709.224312] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 709.224312] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the lease: (returnval){ [ 709.224312] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525599ab-4a17-623a-7256-07ab9aa4f959" [ 709.224312] env[65107]: _type = "HttpNfcLease" [ 709.224312] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 709.225053] env[65107]: INFO nova.compute.manager [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Terminating instance [ 709.241311] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 709.241311] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525599ab-4a17-623a-7256-07ab9aa4f959" [ 709.241311] env[65107]: _type = "HttpNfcLease" [ 709.241311] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 709.241960] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 709.241960] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525599ab-4a17-623a-7256-07ab9aa4f959" [ 709.241960] env[65107]: _type = "HttpNfcLease" [ 709.241960] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 709.245256] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048001b8-bc4d-4712-a0de-6e356a17e9c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.258343] env[65107]: DEBUG nova.compute.manager [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 709.258491] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.259109] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ba93c3-5b0b-175a-1e40-4552838c4a6a/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 709.259664] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ba93c3-5b0b-175a-1e40-4552838c4a6a/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 709.261680] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565d3f7a-5a10-4222-b58e-511f461ae0ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.323580] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 709.331873] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 709.333838] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4773c892-f6ba-4191-aa75-03f21c6e167e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.342923] env[65107]: DEBUG oslo_vmware.api [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 709.342923] env[65107]: value = "task-5102482" [ 709.342923] env[65107]: _type = "Task" [ 709.342923] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.355526] env[65107]: DEBUG nova.network.neutron [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Successfully created port: 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 709.364177] env[65107]: DEBUG oslo_vmware.api [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.399186] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6d1e1bcf-6413-4a1f-ae2f-8818d5021fe4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.533369] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cc3d81-164f-40f7-98bc-bc5597270bb0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.543285] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8357686a-4824-4868-9e3e-8edaa5780209 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.589510] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d60a50-6c44-449d-9bef-6ab9ab019e69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.597812] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102480, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.602678] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d691b86c-5fc4-47cc-9bc5-0c84fd5d134f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.617425] env[65107]: DEBUG nova.compute.provider_tree [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.692361] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f02f74-d53a-c234-c3d7-e446a81f4f85, 'name': SearchDatastore_Task, 'duration_secs': 0.01192} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.692701] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.693015] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.693369] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49c7c4ee-deff-40d5-a273-ec65cdbe910b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.703610] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 709.703610] env[65107]: value = "task-5102483" [ 709.703610] env[65107]: _type = "Task" [ 709.703610] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.715311] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102483, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.737440] env[65107]: DEBUG nova.compute.manager [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 709.737717] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.738753] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f097b4-3bd5-4226-afbc-fa212155f2de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.749197] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 709.749267] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0dd259b-c263-45e4-aee1-d8786b18d589 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.757067] env[65107]: DEBUG oslo_vmware.api [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 709.757067] env[65107]: value = "task-5102484" [ 709.757067] env[65107]: _type = "Task" [ 709.757067] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.769067] env[65107]: DEBUG oslo_vmware.api [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102484, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.855746] env[65107]: DEBUG oslo_vmware.api [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102482, 'name': PowerOffVM_Task, 'duration_secs': 0.457255} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.856167] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 709.856512] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 709.857015] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-301a6396-00f8-4098-9387-7c95e68ca52f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.941844] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 709.942382] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 709.944049] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleting the datastore file [datastore1] 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 709.944499] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a8e0bdc-19e8-40a7-9dce-ceff913fcf61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.959269] env[65107]: DEBUG oslo_vmware.api [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 709.959269] env[65107]: value = "task-5102486" [ 709.959269] env[65107]: _type = "Task" [ 709.959269] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.971753] env[65107]: DEBUG oslo_vmware.api [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.100411] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102480, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.120760] env[65107]: DEBUG nova.scheduler.client.report [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 710.219039] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102483, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.274934] env[65107]: DEBUG oslo_vmware.api [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102484, 'name': PowerOffVM_Task, 'duration_secs': 0.223408} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.275595] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 710.276118] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 710.276520] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43b2c18f-28cb-4f5f-9788-a75a4c69b2d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.341020] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 710.364323] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 710.364323] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 710.364323] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleting the datastore file [datastore1] 07b83b81-9c65-4180-b208-c96525b78d8a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 710.364892] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c100c37d-f796-45ba-a2d0-5b893ed22ec0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.369892] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 710.370400] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 710.370695] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 710.370927] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 710.371430] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 710.371738] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 710.372085] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.372353] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 710.372625] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 710.372886] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 710.373232] env[65107]: DEBUG nova.virt.hardware [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 710.374460] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0e1ce1-f967-453f-9e2e-97a38c8452c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.389492] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd07c868-ae83-4434-9ed2-31e30b3a3b89 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.396512] env[65107]: DEBUG oslo_vmware.api [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for the task: (returnval){ [ 710.396512] env[65107]: value = "task-5102488" [ 710.396512] env[65107]: _type = "Task" [ 710.396512] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.419818] env[65107]: DEBUG oslo_vmware.api [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.472306] env[65107]: DEBUG oslo_vmware.api [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.386682} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.472919] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 710.473018] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 710.473296] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.473856] env[65107]: INFO nova.compute.manager [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Took 1.22 seconds to destroy the instance on the hypervisor. [ 710.474542] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 710.474542] env[65107]: DEBUG nova.compute.manager [-] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 710.475559] env[65107]: DEBUG nova.network.neutron [-] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 710.475559] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.476351] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.476351] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.596921] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102480, 'name': CloneVM_Task, 'duration_secs': 1.329055} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.597231] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Created linked-clone VM from snapshot [ 710.598011] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503af7d1-b670-4833-b366-c93aab004853 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.607264] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Uploading image 7aebf02a-b63e-47e4-96fe-9da51bed3b49 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 710.627883] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.628444] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 710.632995] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 710.632995] env[65107]: value = "vm-992668" [ 710.632995] env[65107]: _type = "VirtualMachine" [ 710.632995] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 710.633330] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.686s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.635312] env[65107]: INFO nova.compute.claims [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.637823] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-939ca8db-60a4-410d-9715-04c2a9f48a2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.648047] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lease: (returnval){ [ 710.648047] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5212b1dd-9617-7b63-4cf5-37bc005f73cf" [ 710.648047] env[65107]: _type = "HttpNfcLease" [ 710.648047] env[65107]: } obtained for exporting VM: (result){ [ 710.648047] env[65107]: value = "vm-992668" [ 710.648047] env[65107]: _type = "VirtualMachine" [ 710.648047] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 710.648047] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the lease: (returnval){ [ 710.648047] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5212b1dd-9617-7b63-4cf5-37bc005f73cf" [ 710.648047] env[65107]: _type = "HttpNfcLease" [ 710.648047] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 710.657131] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 710.657131] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5212b1dd-9617-7b63-4cf5-37bc005f73cf" [ 710.657131] env[65107]: _type = "HttpNfcLease" [ 710.657131] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 710.719845] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102483, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52063} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.722619] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 710.722619] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.722619] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-258a3eca-cfc9-4c75-8c9b-a59d87870bd6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.730804] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 710.730804] env[65107]: value = "task-5102490" [ 710.730804] env[65107]: _type = "Task" [ 710.730804] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.743037] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.750136] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.906248] env[65107]: DEBUG oslo_vmware.api [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Task: {'id': task-5102488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159392} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.906567] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 710.906774] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 710.907044] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.907273] env[65107]: INFO nova.compute.manager [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Took 1.17 seconds to destroy the instance on the hypervisor. [ 710.907535] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 710.907823] env[65107]: DEBUG nova.compute.manager [-] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 710.907992] env[65107]: DEBUG nova.network.neutron [-] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 710.908291] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.908850] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.909168] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.046670] env[65107]: DEBUG nova.network.neutron [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Successfully updated port: 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 711.145688] env[65107]: DEBUG nova.compute.utils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 711.148661] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 711.149370] env[65107]: DEBUG nova.network.neutron [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 711.149862] env[65107]: WARNING neutronclient.v2_0.client [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.150454] env[65107]: WARNING neutronclient.v2_0.client [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.151926] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.152699] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.174026] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 711.174026] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5212b1dd-9617-7b63-4cf5-37bc005f73cf" [ 711.174026] env[65107]: _type = "HttpNfcLease" [ 711.174026] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 711.174026] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 711.174026] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5212b1dd-9617-7b63-4cf5-37bc005f73cf" [ 711.174026] env[65107]: _type = "HttpNfcLease" [ 711.174026] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 711.174026] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208dfa6c-2b09-4039-b503-ea26605746d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.183504] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9da15-647f-7293-90bd-4bde2bb82bf5/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 711.184268] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9da15-647f-7293-90bd-4bde2bb82bf5/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 711.260059] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072499} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.260566] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 711.261914] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe826a5-d8ac-470a-b1a1-c7a61b8724f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.289466] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 711.292477] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65a5c39d-a42e-4263-be3b-c86bccaf933f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.308524] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cda21b3e-3ed7-41db-aaf3-1148f71597f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.320083] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 711.320083] env[65107]: value = "task-5102491" [ 711.320083] env[65107]: _type = "Task" [ 711.320083] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.330108] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102491, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.498244] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.550739] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.550739] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.550739] env[65107]: DEBUG nova.network.neutron [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 711.604403] env[65107]: DEBUG nova.policy [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ccab6477dd44878f78a9d6428d3cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4778fe9152224fd29d1f6220a19b5a36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 711.649256] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 711.836188] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102491, 'name': ReconfigVM_Task, 'duration_secs': 0.296352} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.839971] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Reconfigured VM instance instance-00000002 to attach disk [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 711.840788] env[65107]: DEBUG nova.network.neutron [-] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 711.842973] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17666c37-27ea-4444-a685-f2f5cd6d6b92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.854248] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 711.854248] env[65107]: value = "task-5102492" [ 711.854248] env[65107]: _type = "Task" [ 711.854248] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.878109] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102492, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.052863] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.053799] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.192369] env[65107]: DEBUG nova.network.neutron [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Successfully created port: abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 712.293921] env[65107]: DEBUG nova.network.neutron [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 712.346273] env[65107]: INFO nova.compute.manager [-] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Took 1.87 seconds to deallocate network for instance. [ 712.386824] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102492, 'name': Rename_Task, 'duration_secs': 0.194247} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.388683] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 712.388683] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7539c47-abfd-4574-a0d6-7e12ab1651d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.403292] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 712.403292] env[65107]: value = "task-5102493" [ 712.403292] env[65107]: _type = "Task" [ 712.403292] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.419512] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102493, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.449955] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7496a5-af47-49b7-8724-7da322a26dc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.463788] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ac8e05-f3a0-4a6c-88a7-bcdc7f4cdedc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.508368] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989fdd1e-17ed-47bb-9334-4ebc5a1928f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.519514] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126a1579-476d-4064-a3d8-467fada28fec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.536921] env[65107]: DEBUG nova.network.neutron [-] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 712.538544] env[65107]: DEBUG nova.compute.provider_tree [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.666418] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 712.868248] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.916166] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102493, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.043101] env[65107]: INFO nova.compute.manager [-] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Took 2.13 seconds to deallocate network for instance. [ 713.048688] env[65107]: DEBUG nova.scheduler.client.report [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.420393] env[65107]: DEBUG oslo_vmware.api [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102493, 'name': PowerOnVM_Task, 'duration_secs': 0.658033} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.421284] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 713.421799] env[65107]: DEBUG nova.compute.manager [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 713.422858] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4599fa0-d9bb-4029-827e-05f84499a26d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.458276] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.458276] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.562571] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.928s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.563304] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 713.569781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.570078] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.572s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.572794] env[65107]: INFO nova.compute.claims [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.945844] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.032480] env[65107]: DEBUG nova.network.neutron [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Successfully updated port: abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 714.049031] env[65107]: WARNING neutronclient.v2_0.client [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.049760] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.050133] env[65107]: WARNING openstack [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.078908] env[65107]: DEBUG nova.compute.utils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 714.085565] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 714.085565] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 714.085828] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.086053] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.086666] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.087146] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.232683] env[65107]: DEBUG nova.policy [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '796101bf412844bf9c547a0d4376d8d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c347db6748d044e8842e336a7f25c815', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 714.380726] env[65107]: DEBUG nova.network.neutron [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.504426] env[65107]: DEBUG nova.compute.manager [req-6bdacda4-170f-4821-aa86-bcf21d2da0af req-5db642f5-5033-4d3c-b306-ce1897409f00 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-vif-plugged-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 714.504695] env[65107]: DEBUG oslo_concurrency.lockutils [req-6bdacda4-170f-4821-aa86-bcf21d2da0af req-5db642f5-5033-4d3c-b306-ce1897409f00 service nova] Acquiring lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.504929] env[65107]: DEBUG oslo_concurrency.lockutils [req-6bdacda4-170f-4821-aa86-bcf21d2da0af req-5db642f5-5033-4d3c-b306-ce1897409f00 service nova] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.505115] env[65107]: DEBUG oslo_concurrency.lockutils [req-6bdacda4-170f-4821-aa86-bcf21d2da0af req-5db642f5-5033-4d3c-b306-ce1897409f00 service nova] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.505719] env[65107]: DEBUG nova.compute.manager [req-6bdacda4-170f-4821-aa86-bcf21d2da0af req-5db642f5-5033-4d3c-b306-ce1897409f00 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] No waiting events found dispatching network-vif-plugged-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 714.505719] env[65107]: WARNING nova.compute.manager [req-6bdacda4-170f-4821-aa86-bcf21d2da0af req-5db642f5-5033-4d3c-b306-ce1897409f00 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received unexpected event network-vif-plugged-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d for instance with vm_state building and task_state spawning. [ 714.535281] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.535466] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.535665] env[65107]: DEBUG nova.network.neutron [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 714.586187] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 714.688043] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e4dcdf-9547-4ed1-879c-9607070e65d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.697009] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95551da-866a-4fdd-ab51-37b0b1645d66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.737229] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b49d34b-c6e5-4f2a-8780-430b8ae85864 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.746330] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46868fe-bf19-44c8-ae4c-102c2ead99fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.762541] env[65107]: DEBUG nova.compute.provider_tree [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.764881] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Successfully created port: 53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 714.887636] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.888112] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Instance network_info: |[{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 714.888665] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:ad:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee4b2432-c393-4e50-ae0e-b5e12bad37db', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ec2fffa-a684-4c3d-8ead-2c8b74d6252d', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.900173] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Creating folder: Project (58588fe2a11f4ee0aab313903b211fa8). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.900655] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa32814a-6830-43a0-9f5b-928d88af2ba5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.914939] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Created folder: Project (58588fe2a11f4ee0aab313903b211fa8) in parent group-v992574. [ 714.914939] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Creating folder: Instances. Parent ref: group-v992669. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.914939] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-769019f0-1b18-446e-81bc-f598b15e3308 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.927479] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Created folder: Instances in parent group-v992669. [ 714.927823] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 714.928064] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 714.928469] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f09be1e-15a3-4771-a476-26b1a96ebd01 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.951895] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.951895] env[65107]: value = "task-5102496" [ 714.951895] env[65107]: _type = "Task" [ 714.951895] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.961207] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102496, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.041823] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.042498] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.093097] env[65107]: DEBUG nova.network.neutron [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 715.181177] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.181693] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.266068] env[65107]: WARNING neutronclient.v2_0.client [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 715.266970] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.267416] env[65107]: WARNING openstack [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.277527] env[65107]: DEBUG nova.scheduler.client.report [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.463290] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102496, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.603910] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 715.643278] env[65107]: DEBUG nova.network.neutron [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Updating instance_info_cache with network_info: [{"id": "abf46aae-348e-4b4b-acfe-35830ba162e1", "address": "fa:16:3e:af:30:b1", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabf46aae-34", "ovs_interfaceid": "abf46aae-348e-4b4b-acfe-35830ba162e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 715.782954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 715.783459] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 715.786667] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.240s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 715.788422] env[65107]: INFO nova.compute.claims [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.964229] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102496, 'name': CreateVM_Task, 'duration_secs': 0.52829} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.964550] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 715.965190] env[65107]: WARNING neutronclient.v2_0.client [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 715.965724] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.965924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 715.966420] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 715.966698] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f1a91e8-423a-4759-9c2a-4097f32d7e05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.973436] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 715.973436] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5297dfa7-7171-7ac9-823e-fee1d4733bda" [ 715.973436] env[65107]: _type = "Task" [ 715.973436] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.984360] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5297dfa7-7171-7ac9-823e-fee1d4733bda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.145911] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.146448] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Instance network_info: |[{"id": "abf46aae-348e-4b4b-acfe-35830ba162e1", "address": "fa:16:3e:af:30:b1", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabf46aae-34", "ovs_interfaceid": "abf46aae-348e-4b4b-acfe-35830ba162e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 716.294079] env[65107]: DEBUG nova.compute.utils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 716.298535] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 716.299585] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 716.299585] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.299585] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.299934] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.300329] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.441146] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Successfully updated port: 53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 716.488734] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5297dfa7-7171-7ac9-823e-fee1d4733bda, 'name': SearchDatastore_Task, 'duration_secs': 0.010173} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.495407] env[65107]: DEBUG nova.policy [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '796101bf412844bf9c547a0d4376d8d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c347db6748d044e8842e336a7f25c815', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 716.497512] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.497894] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 716.498341] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.498530] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.498758] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.499874] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f31b147e-1320-4218-b626-e434696b9e09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.511251] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.511469] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 716.516200] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-819947b4-446a-45c5-9612-588f61d9b1aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.523249] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 716.523249] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ddde97-ad51-b1a1-61a4-5ddda08f30e9" [ 716.523249] env[65107]: _type = "Task" [ 716.523249] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.535478] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ddde97-ad51-b1a1-61a4-5ddda08f30e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.799605] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 716.821756] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Successfully created port: 7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 716.892239] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e35735c-c4c4-4acc-a697-80a52a2228c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.903707] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7422290d-844e-44df-ba15-60f8724a3a07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.944792] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca6db8a-7f4c-41fc-8d94-10bddfa8f30e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.948236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "refresh_cache-3948dcdd-e7c1-40b6-a3ec-97741940a0fb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.948320] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "refresh_cache-3948dcdd-e7c1-40b6-a3ec-97741940a0fb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.948503] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 716.957239] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d1216f-73cb-4b38-8028-c297c999f061 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.976511] env[65107]: DEBUG nova.compute.provider_tree [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.034734] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ddde97-ad51-b1a1-61a4-5ddda08f30e9, 'name': SearchDatastore_Task, 'duration_secs': 0.012312} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.035732] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d90b1980-f4ec-428e-9f27-933ea95f1f2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.042771] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 717.042771] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cee72-6a1b-8ee3-4070-508379c2bd77" [ 717.042771] env[65107]: _type = "Task" [ 717.042771] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.053723] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cee72-6a1b-8ee3-4070-508379c2bd77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.452013] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.452550] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.481081] env[65107]: DEBUG nova.scheduler.client.report [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 717.555609] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cee72-6a1b-8ee3-4070-508379c2bd77, 'name': SearchDatastore_Task, 'duration_secs': 0.013168} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.555910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.556196] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ef417f32-8c0e-4a62-9078-cb2bf95b85f8/ef417f32-8c0e-4a62-9078-cb2bf95b85f8.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 717.556472] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fa45457-221d-490b-86c7-278a4438a94c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.565328] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 717.565328] env[65107]: value = "task-5102497" [ 717.565328] env[65107]: _type = "Task" [ 717.565328] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.574736] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.590685] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 717.629851] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.630014] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.707045] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.707045] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.707045] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.802306] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Updating instance_info_cache with network_info: [{"id": "53ef6db1-9d79-43ac-ad0a-c794a9fed04d", "address": "fa:16:3e:1a:2d:5f", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53ef6db1-9d", "ovs_interfaceid": "53ef6db1-9d79-43ac-ad0a-c794a9fed04d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 717.814570] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 717.986214] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.986905] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 717.990091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.984s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.991845] env[65107]: INFO nova.compute.claims [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.076221] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102497, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.305383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "refresh_cache-3948dcdd-e7c1-40b6-a3ec-97741940a0fb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.306055] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Instance network_info: |[{"id": "53ef6db1-9d79-43ac-ad0a-c794a9fed04d", "address": "fa:16:3e:1a:2d:5f", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53ef6db1-9d", "ovs_interfaceid": "53ef6db1-9d79-43ac-ad0a-c794a9fed04d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 718.440360] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Successfully updated port: 7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 718.498817] env[65107]: DEBUG nova.compute.utils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 718.504301] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 718.504301] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 718.504301] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.504694] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.505862] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.505862] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.566287] env[65107]: DEBUG nova.policy [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '796101bf412844bf9c547a0d4376d8d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c347db6748d044e8842e336a7f25c815', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 718.577525] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527817} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.577784] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ef417f32-8c0e-4a62-9078-cb2bf95b85f8/ef417f32-8c0e-4a62-9078-cb2bf95b85f8.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 718.577993] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.578650] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5024bf70-f46b-47dc-b97e-a12b465b61ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.585667] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 718.585667] env[65107]: value = "task-5102498" [ 718.585667] env[65107]: _type = "Task" [ 718.585667] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.596225] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102498, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.664688] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 718.667725] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.667868] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 718.668313] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.668528] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 718.668707] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 718.669097] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.669194] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 718.669305] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 718.669466] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 718.669636] env[65107]: DEBUG nova.virt.hardware [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 718.680018] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa3ebc9-2164-4acb-a8b4-18f2d231458b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.692782] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 718.692782] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.692782] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 718.692962] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.692962] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 718.692962] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 718.692962] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.692962] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 718.693126] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 718.693126] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 718.693126] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 718.694692] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008bdc20-98f3-4022-98b4-88fadc8470fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.700879] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3241d9-5607-4acf-8130-c7480ca76701 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.711298] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 718.711530] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.712886] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 718.712886] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.712886] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 718.712886] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 718.712886] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.713098] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 718.713098] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 718.713202] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 718.713363] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 718.723316] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c5b33e-0071-4987-ba2a-13fc51a66994 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.729536] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b58efee-5946-4639-a2c1-f62b48baff46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.733028] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:30:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'abf46aae-348e-4b4b-acfe-35830ba162e1', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.740409] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 718.741099] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 718.742152] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e606a89-e18e-4ec1-b2fb-bed7befbb7f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.769028] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ba93c3-5b0b-175a-1e40-4552838c4a6a/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 718.774807] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:2d:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca50cd14-9e1f-4d74-a066-e5a45ba0ce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53ef6db1-9d79-43ac-ad0a-c794a9fed04d', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.779891] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Creating folder: Project (c347db6748d044e8842e336a7f25c815). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.781207] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3831dcb5-bf88-4a73-a59e-461ec0416a27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.785332] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-42c75c27-dd53-4562-a208-e3a6da67f07d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.787531] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab543ab8-f017-4943-a9b4-3a42451e97b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.796025] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.796025] env[65107]: value = "task-5102499" [ 718.796025] env[65107]: _type = "Task" [ 718.796025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.799752] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ba93c3-5b0b-175a-1e40-4552838c4a6a/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 718.799935] env[65107]: ERROR oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ba93c3-5b0b-175a-1e40-4552838c4a6a/disk-0.vmdk due to incomplete transfer. [ 718.808357] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1a52b29e-21d4-47a3-9f80-7eb4d8a5c27e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.815266] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Created folder: Project (c347db6748d044e8842e336a7f25c815) in parent group-v992574. [ 718.818095] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Creating folder: Instances. Parent ref: group-v992672. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.818095] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102499, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.818095] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa39c40d-06ec-46d1-b8a5-1631646517b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.831438] env[65107]: DEBUG oslo_vmware.rw_handles [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ba93c3-5b0b-175a-1e40-4552838c4a6a/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 718.831660] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Uploaded image 39f72a8a-740a-4530-923e-597c889d6992 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 718.834300] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 718.836015] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-19e613ba-d5a2-4d36-845a-b1aab8802030 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.838143] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Created folder: Instances in parent group-v992672. [ 718.838400] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 718.838622] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 718.839226] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd5996e2-c918-4de4-b342-d1a7154206c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.856740] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 718.856740] env[65107]: value = "task-5102502" [ 718.856740] env[65107]: _type = "Task" [ 718.856740] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.861719] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Successfully created port: 52873450-f314-4110-a316-7380569e0afe {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 718.866655] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.866655] env[65107]: value = "task-5102503" [ 718.866655] env[65107]: _type = "Task" [ 718.866655] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.873872] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102502, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.879513] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102503, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.945389] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "refresh_cache-c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.945887] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "refresh_cache-c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.945887] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 718.965958] env[65107]: DEBUG nova.compute.manager [req-c9275775-dea2-432b-b45d-dce68e92c525 req-ca91f881-d53a-4b32-ac1e-723a90a84efa service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Received event network-vif-plugged-abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 718.966352] env[65107]: DEBUG oslo_concurrency.lockutils [req-c9275775-dea2-432b-b45d-dce68e92c525 req-ca91f881-d53a-4b32-ac1e-723a90a84efa service nova] Acquiring lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.966756] env[65107]: DEBUG oslo_concurrency.lockutils [req-c9275775-dea2-432b-b45d-dce68e92c525 req-ca91f881-d53a-4b32-ac1e-723a90a84efa service nova] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.967083] env[65107]: DEBUG oslo_concurrency.lockutils [req-c9275775-dea2-432b-b45d-dce68e92c525 req-ca91f881-d53a-4b32-ac1e-723a90a84efa service nova] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.968142] env[65107]: DEBUG nova.compute.manager [req-c9275775-dea2-432b-b45d-dce68e92c525 req-ca91f881-d53a-4b32-ac1e-723a90a84efa service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] No waiting events found dispatching network-vif-plugged-abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 718.968351] env[65107]: WARNING nova.compute.manager [req-c9275775-dea2-432b-b45d-dce68e92c525 req-ca91f881-d53a-4b32-ac1e-723a90a84efa service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Received unexpected event network-vif-plugged-abf46aae-348e-4b4b-acfe-35830ba162e1 for instance with vm_state building and task_state spawning. [ 719.007191] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 719.097307] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102498, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069098} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.100239] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.101667] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff323c9-2f8b-47e2-9613-5dad6d894857 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.125462] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] ef417f32-8c0e-4a62-9078-cb2bf95b85f8/ef417f32-8c0e-4a62-9078-cb2bf95b85f8.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.129761] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62e457af-ef6d-4a91-975d-e72ec25995ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.150806] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 719.150806] env[65107]: value = "task-5102504" [ 719.150806] env[65107]: _type = "Task" [ 719.150806] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.164956] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102504, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.309807] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102499, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.377927] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102502, 'name': Destroy_Task, 'duration_secs': 0.402776} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.378494] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Destroyed the VM [ 719.378877] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 719.379725] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d8890344-cc47-42c0-8084-abcaea379c77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.387482] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102503, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.394413] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 719.394413] env[65107]: value = "task-5102505" [ 719.394413] env[65107]: _type = "Task" [ 719.394413] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.405010] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102505, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.451287] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.451685] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.503248] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 719.565665] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.566100] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.635751] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241832e8-6e0d-424c-9c3d-ededfc572874 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.650690] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa06a96-d860-4d0d-92fe-9b7e196d78ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.664605] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102504, 'name': ReconfigVM_Task, 'duration_secs': 0.367423} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.691603] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Reconfigured VM instance instance-0000001f to attach disk [datastore1] ef417f32-8c0e-4a62-9078-cb2bf95b85f8/ef417f32-8c0e-4a62-9078-cb2bf95b85f8.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.692595] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-822eec92-a514-49c6-802b-280e90414053 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.695059] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f180a8d-064a-4872-a96f-c8cab78ef51f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.707122] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f142e6e7-39e3-48a8-a56f-a4e0fee9485a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.710330] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 719.710330] env[65107]: value = "task-5102506" [ 719.710330] env[65107]: _type = "Task" [ 719.710330] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.722859] env[65107]: DEBUG nova.compute.provider_tree [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.725773] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.726467] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.726846] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.745746] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102506, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.813074] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102499, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.861238] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Updating instance_info_cache with network_info: [{"id": "7b78f034-800f-47c1-9c6f-7e94df8703b1", "address": "fa:16:3e:39:5f:1a", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b78f034-80", "ovs_interfaceid": "7b78f034-800f-47c1-9c6f-7e94df8703b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 719.882299] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102503, 'name': CreateVM_Task, 'duration_secs': 0.571977} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.883200] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.883795] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.884245] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.884453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.884789] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 719.885443] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afb03d3b-ba02-4e64-ba95-1f14006f9747 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.894701] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 719.894701] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524cfa2c-a7bf-e8e6-5d0d-74fac66a7b15" [ 719.894701] env[65107]: _type = "Task" [ 719.894701] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.913299] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524cfa2c-a7bf-e8e6-5d0d-74fac66a7b15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.913570] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102505, 'name': RemoveSnapshot_Task} progress is 60%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.022140] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 720.055280] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 720.055604] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 720.055814] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 720.056029] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 720.056441] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 720.056441] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 720.056545] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.056679] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 720.057030] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 720.057030] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 720.057226] env[65107]: DEBUG nova.virt.hardware [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 720.058350] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd555c56-774f-43ec-a8f4-bab104beef5c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.067672] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bafb1dc-09cf-4436-8a0f-b0c09036186f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.161810] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.161810] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.220147] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102506, 'name': Rename_Task, 'duration_secs': 0.159642} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.221596] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 720.221596] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-694935d9-76b9-4331-b9c7-c921d6b2fed7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.227964] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 720.227964] env[65107]: value = "task-5102507" [ 720.227964] env[65107]: _type = "Task" [ 720.227964] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.237371] env[65107]: DEBUG nova.scheduler.client.report [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.240814] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.308272] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102499, 'name': CreateVM_Task, 'duration_secs': 1.23249} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.308553] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 720.309208] env[65107]: WARNING neutronclient.v2_0.client [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.309752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.365680] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "refresh_cache-c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.369504] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Instance network_info: |[{"id": "7b78f034-800f-47c1-9c6f-7e94df8703b1", "address": "fa:16:3e:39:5f:1a", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b78f034-80", "ovs_interfaceid": "7b78f034-800f-47c1-9c6f-7e94df8703b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 720.369847] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:5f:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca50cd14-9e1f-4d74-a066-e5a45ba0ce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b78f034-800f-47c1-9c6f-7e94df8703b1', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.378326] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 720.379596] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 720.379596] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4910e660-5dd4-4fab-90d6-106f55c70541 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.406214] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.406214] env[65107]: value = "task-5102508" [ 720.406214] env[65107]: _type = "Task" [ 720.406214] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.409674] env[65107]: DEBUG oslo_vmware.api [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102505, 'name': RemoveSnapshot_Task, 'duration_secs': 0.650394} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.418475] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 720.418879] env[65107]: INFO nova.compute.manager [None req-abab80a0-ac3d-4264-bb58-56f23c7c5875 tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Took 14.94 seconds to snapshot the instance on the hypervisor. [ 720.423309] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524cfa2c-a7bf-e8e6-5d0d-74fac66a7b15, 'name': SearchDatastore_Task, 'duration_secs': 0.020588} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.424995] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.425479] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.426105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.426308] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.426539] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 720.427379] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.427765] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 720.428080] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a09a2300-d01b-4005-a6a3-f1cf8bcb2173 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.435587] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-825da3f6-f62f-4c00-9b3c-4d550ddd6f92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.437607] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102508, 'name': CreateVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.448914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "6563cbda-5fd3-4640-9e9a-95904f3f8335" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.449277] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.449604] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.449826] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 720.450801] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 720.450801] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241ee5a-4092-6bec-8755-71c55101c4a2" [ 720.450801] env[65107]: _type = "Task" [ 720.450801] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.451541] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb5d1054-78e2-49c2-ac83-acc629bf603d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.462967] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 720.462967] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528d650b-8b2b-750a-3257-770f6c7583a2" [ 720.462967] env[65107]: _type = "Task" [ 720.462967] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.467052] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241ee5a-4092-6bec-8755-71c55101c4a2, 'name': SearchDatastore_Task, 'duration_secs': 0.01091} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.471316] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.471670] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.471943] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.481512] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528d650b-8b2b-750a-3257-770f6c7583a2, 'name': SearchDatastore_Task, 'duration_secs': 0.01151} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.482395] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b22b888f-5ad8-43d3-96ac-b698795eb3de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.490245] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 720.490245] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f6bbda-ed64-7dfe-40f5-416c89697905" [ 720.490245] env[65107]: _type = "Task" [ 720.490245] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.502556] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f6bbda-ed64-7dfe-40f5-416c89697905, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.545281] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Successfully updated port: 52873450-f314-4110-a316-7380569e0afe {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 720.656945] env[65107]: DEBUG nova.compute.manager [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 720.657336] env[65107]: DEBUG nova.compute.manager [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing instance network info cache due to event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 720.657805] env[65107]: DEBUG oslo_concurrency.lockutils [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Acquiring lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.658083] env[65107]: DEBUG oslo_concurrency.lockutils [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Acquired lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.658224] env[65107]: DEBUG nova.network.neutron [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 720.738875] env[65107]: DEBUG oslo_vmware.api [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102507, 'name': PowerOnVM_Task, 'duration_secs': 0.507127} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.739166] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 720.739358] env[65107]: INFO nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Took 10.40 seconds to spawn the instance on the hypervisor. [ 720.739824] env[65107]: DEBUG nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 720.740351] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9885cd38-2cc2-4835-a547-1f5528860eb6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.743755] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.754s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.744213] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 720.747369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.340s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.747457] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.749557] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.802s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.751323] env[65107]: INFO nova.compute.claims [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.786980] env[65107]: INFO nova.scheduler.client.report [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Deleted allocations for instance ccd68268-d0fc-406e-896f-c61b114dd75d [ 720.921278] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102508, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.003883] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f6bbda-ed64-7dfe-40f5-416c89697905, 'name': SearchDatastore_Task, 'duration_secs': 0.011692} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.004386] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.004871] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 3948dcdd-e7c1-40b6-a3ec-97741940a0fb/3948dcdd-e7c1-40b6-a3ec-97741940a0fb.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 721.005318] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.006031] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.006386] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-941bd5b2-2d13-4f03-a5ed-cb31832f2464 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.010373] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0154b774-6b0a-4335-b62d-8dd8f8817486 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.019726] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 721.019726] env[65107]: value = "task-5102509" [ 721.019726] env[65107]: _type = "Task" [ 721.019726] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.024916] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.024916] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 721.025933] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4e17190-14ad-4084-82b2-e5d7cc841d64 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.031949] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102509, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.036214] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 721.036214] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521801dd-a1c0-3600-f3a8-4958a6992650" [ 721.036214] env[65107]: _type = "Task" [ 721.036214] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.046423] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521801dd-a1c0-3600-f3a8-4958a6992650, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.048482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "refresh_cache-90706f49-af87-4038-ae76-02a95e4a12e6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.048482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "refresh_cache-90706f49-af87-4038-ae76-02a95e4a12e6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.048482] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 721.161483] env[65107]: WARNING neutronclient.v2_0.client [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.163339] env[65107]: WARNING openstack [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.163943] env[65107]: WARNING openstack [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.262898] env[65107]: DEBUG nova.compute.utils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 721.274087] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 721.275537] env[65107]: DEBUG nova.network.neutron [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 721.275537] env[65107]: WARNING neutronclient.v2_0.client [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.275537] env[65107]: WARNING neutronclient.v2_0.client [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.277900] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.277900] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.295601] env[65107]: INFO nova.compute.manager [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Took 41.41 seconds to build instance. [ 721.297328] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4aa95487-bd6a-4953-886c-2642f02aa176 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016 tempest-FloatingIPsAssociationNegativeTestJSON-1021891016-project-member] Lock "ccd68268-d0fc-406e-896f-c61b114dd75d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.465s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.313690] env[65107]: WARNING openstack [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.314135] env[65107]: WARNING openstack [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.390376] env[65107]: WARNING neutronclient.v2_0.client [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.391153] env[65107]: WARNING openstack [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.391505] env[65107]: WARNING openstack [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.406709] env[65107]: DEBUG nova.policy [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ea13ac95bd94c39851197d5d04f3492', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '847cf5be309646b59ed21971ea28f0a2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 721.424519] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102508, 'name': CreateVM_Task, 'duration_secs': 0.540612} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.424766] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 721.425505] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.425805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.425933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.426408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 721.426817] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cdfbb91-b739-4026-a4ee-a306a262e049 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.433506] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 721.433506] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52276b3f-01c9-e89c-0cc1-612b26d3636d" [ 721.433506] env[65107]: _type = "Task" [ 721.433506] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.442043] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9da15-647f-7293-90bd-4bde2bb82bf5/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 721.443046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d579c5-1bc2-4140-b869-3f141dd2134b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.451278] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52276b3f-01c9-e89c-0cc1-612b26d3636d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.460753] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9da15-647f-7293-90bd-4bde2bb82bf5/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 721.460936] env[65107]: ERROR oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9da15-647f-7293-90bd-4bde2bb82bf5/disk-0.vmdk due to incomplete transfer. [ 721.461209] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b99a187e-c230-48c0-a6cf-0801a0a07338 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.469968] env[65107]: DEBUG oslo_vmware.rw_handles [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9da15-647f-7293-90bd-4bde2bb82bf5/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 721.470301] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Uploaded image 7aebf02a-b63e-47e4-96fe-9da51bed3b49 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 721.471752] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 721.472052] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8f18a4b4-de9d-4468-aba9-d5c9aebb9580 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.479998] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 721.479998] env[65107]: value = "task-5102510" [ 721.479998] env[65107]: _type = "Task" [ 721.479998] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.490036] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102510, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.530722] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102509, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.547795] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521801dd-a1c0-3600-f3a8-4958a6992650, 'name': SearchDatastore_Task, 'duration_secs': 0.01349} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.549042] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb69a6fa-cfe1-486e-b31d-1c4bd4b6a6e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.552022] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.552436] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.564923] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 721.564923] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849194-524c-f45f-a02f-5034ef3cc79b" [ 721.564923] env[65107]: _type = "Task" [ 721.564923] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.574021] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849194-524c-f45f-a02f-5034ef3cc79b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.724114] env[65107]: DEBUG nova.network.neutron [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updated VIF entry in instance network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 721.724530] env[65107]: DEBUG nova.network.neutron [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.752082] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 721.774469] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 721.799777] env[65107]: DEBUG oslo_concurrency.lockutils [None req-81c7eb61-303f-40d4-9860-61e4e2727bfe tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.936s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.845389] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.845945] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.868787] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Received event network-changed-abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 721.869040] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Refreshing instance network info cache due to event network-changed-abf46aae-348e-4b4b-acfe-35830ba162e1. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 721.870191] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquiring lock "refresh_cache-c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.870574] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquired lock "refresh_cache-c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.870758] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Refreshing network info cache for port abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 721.880454] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.880676] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.946186] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52276b3f-01c9-e89c-0cc1-612b26d3636d, 'name': SearchDatastore_Task, 'duration_secs': 0.064212} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.949818] env[65107]: DEBUG nova.network.neutron [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Successfully created port: 454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 721.951989] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.952252] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.952444] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.954319] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.954706] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.955066] env[65107]: WARNING openstack [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.990306] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102510, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.032076] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102509, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537219} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.032499] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 3948dcdd-e7c1-40b6-a3ec-97741940a0fb/3948dcdd-e7c1-40b6-a3ec-97741940a0fb.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 722.032830] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.036053] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23e322e3-7e76-449f-9f7f-97aaeb714412 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.043496] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 722.043496] env[65107]: value = "task-5102511" [ 722.043496] env[65107]: _type = "Task" [ 722.043496] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.054195] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102511, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.080704] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849194-524c-f45f-a02f-5034ef3cc79b, 'name': SearchDatastore_Task, 'duration_secs': 0.01292} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.081986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.081986] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] c7739dfd-88d6-46ac-b3a0-e7dfe9b00332/c7739dfd-88d6-46ac-b3a0-e7dfe9b00332.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 722.081986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.081986] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 722.082168] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84dc6e0c-8a40-4f8c-a9ab-eb2b81473563 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.089214] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6da92f3-e338-4e40-8ecb-b098ebbbb147 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.103031] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 722.103031] env[65107]: value = "task-5102512" [ 722.103031] env[65107]: _type = "Task" [ 722.103031] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.106815] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 722.107124] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 722.108435] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2ab60a7-cdca-4e30-979c-b9fbbdb0cad9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.115636] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.125244] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 722.125244] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eaa76e-9363-5089-8378-3698a8627c49" [ 722.125244] env[65107]: _type = "Task" [ 722.125244] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.128323] env[65107]: DEBUG nova.network.neutron [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Updating instance_info_cache with network_info: [{"id": "52873450-f314-4110-a316-7380569e0afe", "address": "fa:16:3e:3c:28:f0", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52873450-f3", "ovs_interfaceid": "52873450-f314-4110-a316-7380569e0afe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 722.142253] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eaa76e-9363-5089-8378-3698a8627c49, 'name': SearchDatastore_Task, 'duration_secs': 0.012423} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.145931] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6daff98e-7369-4d43-80c4-9403d7f4e59d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.154418] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 722.154418] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525fde56-2c8b-6ac9-0a3e-353be2076752" [ 722.154418] env[65107]: _type = "Task" [ 722.154418] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.163794] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525fde56-2c8b-6ac9-0a3e-353be2076752, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.231504] env[65107]: DEBUG oslo_concurrency.lockutils [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] Releasing lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.231504] env[65107]: DEBUG nova.compute.manager [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Received event network-vif-deleted-abb39b9b-672f-498e-b98a-38ab04a19143 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 722.231504] env[65107]: DEBUG nova.compute.manager [req-56e5169a-adc2-4521-b492-632a6dca3b07 req-27bb2f62-1a75-4123-a49d-518c68159d1a service nova] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Received event network-vif-deleted-232c23c9-ffb2-48f4-8094-39a295742f87 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 722.302695] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 722.380944] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.381839] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.382055] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.499367] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102510, 'name': Destroy_Task, 'duration_secs': 0.652956} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.504503] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Destroyed the VM [ 722.505435] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 722.508210] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ec627540-291a-4903-b66c-af51baf6fd02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.519394] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 722.519394] env[65107]: value = "task-5102513" [ 722.519394] env[65107]: _type = "Task" [ 722.519394] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.535841] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102513, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.558025] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102511, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073727} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.558025] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.558273] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05b6a30-1c26-4c36-906f-9ab9614aeea6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.567595] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a57af38-810b-43ae-bf1e-995e1b7e15c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.590950] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 3948dcdd-e7c1-40b6-a3ec-97741940a0fb/3948dcdd-e7c1-40b6-a3ec-97741940a0fb.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.591887] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f49731de-4a90-4b2d-94aa-6473442ce33e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.613994] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0cc0af4-d6b7-4049-b6fc-9697151a22a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.620611] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 722.620611] env[65107]: value = "task-5102514" [ 722.620611] env[65107]: _type = "Task" [ 722.620611] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.630215] env[65107]: INFO nova.compute.manager [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Rebuilding instance [ 722.632644] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102512, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.664223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "refresh_cache-90706f49-af87-4038-ae76-02a95e4a12e6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.664223] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Instance network_info: |[{"id": "52873450-f314-4110-a316-7380569e0afe", "address": "fa:16:3e:3c:28:f0", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52873450-f3", "ovs_interfaceid": "52873450-f314-4110-a316-7380569e0afe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 722.669171] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:28:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca50cd14-9e1f-4d74-a066-e5a45ba0ce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52873450-f314-4110-a316-7380569e0afe', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.678220] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 722.681918] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3ca551-427c-4f4c-a309-dc88dfa195a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.684984] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102514, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.687208] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.687691] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d4c76a8-5c0c-4962-87b9-28830664caa7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.713205] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525fde56-2c8b-6ac9-0a3e-353be2076752, 'name': SearchDatastore_Task, 'duration_secs': 0.027418} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.714454] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24b5c2a-6c2a-4db9-bf5e-62607256fbd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.718687] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.719028] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0/c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 722.720589] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6246a997-a873-428f-a62f-1f58451062f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.723080] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.723080] env[65107]: value = "task-5102515" [ 722.723080] env[65107]: _type = "Task" [ 722.723080] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.731930] env[65107]: DEBUG nova.compute.manager [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 722.741885] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a10c122-4fbe-4d98-98f9-f671577bb9e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.744812] env[65107]: DEBUG nova.compute.provider_tree [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.746398] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 722.746398] env[65107]: value = "task-5102516" [ 722.746398] env[65107]: _type = "Task" [ 722.746398] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.755220] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102515, 'name': CreateVM_Task} progress is 15%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.763457] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102516, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.788412] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 722.818019] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 722.818019] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 722.818019] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 722.818235] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 722.818235] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 722.818528] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 722.818906] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.819254] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 722.819595] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 722.820066] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 722.820387] env[65107]: DEBUG nova.virt.hardware [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 722.821446] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc16ba6-070e-4761-a34c-5681dfb33c57 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.825810] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.832069] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d11518b-b07d-489b-9331-1eee693315b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.033335] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102513, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.122836] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58182} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.127197] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] c7739dfd-88d6-46ac-b3a0-e7dfe9b00332/c7739dfd-88d6-46ac-b3a0-e7dfe9b00332.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 723.127474] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.127876] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5da3b4fb-e7a8-4e96-a220-242ff532b82e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.136168] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 723.136168] env[65107]: value = "task-5102517" [ 723.136168] env[65107]: _type = "Task" [ 723.136168] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.140009] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102514, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.153681] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102517, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.236774] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102515, 'name': CreateVM_Task} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.237078] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.237625] env[65107]: WARNING neutronclient.v2_0.client [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.238092] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.238246] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.238482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 723.238747] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d902d237-399e-42a3-8824-bd5d014fc876 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.244079] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.244079] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5269fce6-1223-23f8-4634-48d7ec429049" [ 723.244079] env[65107]: _type = "Task" [ 723.244079] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.262607] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5269fce6-1223-23f8-4634-48d7ec429049, 'name': SearchDatastore_Task, 'duration_secs': 0.009296} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.266328] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.266982] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.266982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.266982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.267240] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.267471] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102516, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47329} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.268438] env[65107]: ERROR nova.scheduler.client.report [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [req-eb8bddf4-3b43-4e78-8fdb-c6feadc93664] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-eb8bddf4-3b43-4e78-8fdb-c6feadc93664"}]} [ 723.268848] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26d35de0-c18c-4434-954d-8d5b0c3dbe88 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.270795] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0/c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 723.271014] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.273922] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fe36ff1-1bef-48f9-a354-ff9a4227b530 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.282110] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.282110] env[65107]: value = "task-5102518" [ 723.282110] env[65107]: _type = "Task" [ 723.282110] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.287607] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.287745] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.289400] env[65107]: DEBUG nova.scheduler.client.report [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 723.292016] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3252b277-bbb7-4b1e-96f0-297423d774dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.299524] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.303334] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.303334] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527239d0-85e4-0b1f-b6d2-35298a1d5fb6" [ 723.303334] env[65107]: _type = "Task" [ 723.303334] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.313114] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527239d0-85e4-0b1f-b6d2-35298a1d5fb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.315531] env[65107]: DEBUG nova.scheduler.client.report [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 723.316113] env[65107]: DEBUG nova.compute.provider_tree [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.327854] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.328675] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.335694] env[65107]: DEBUG nova.scheduler.client.report [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 723.360706] env[65107]: DEBUG nova.scheduler.client.report [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 723.530689] env[65107]: DEBUG oslo_vmware.api [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102513, 'name': RemoveSnapshot_Task, 'duration_secs': 0.896837} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.533632] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 723.533919] env[65107]: INFO nova.compute.manager [None req-a454bc72-aed7-48ec-86c9-499f8bb9e994 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Took 17.08 seconds to snapshot the instance on the hypervisor. [ 723.633763] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102514, 'name': ReconfigVM_Task, 'duration_secs': 0.62901} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.634053] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 3948dcdd-e7c1-40b6-a3ec-97741940a0fb/3948dcdd-e7c1-40b6-a3ec-97741940a0fb.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.634885] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72760130-f989-45f6-809f-f8c53b973b2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.644045] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.644045] env[65107]: value = "task-5102519" [ 723.644045] env[65107]: _type = "Task" [ 723.644045] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.654226] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102517, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110894} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.654881] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.655724] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458a37eb-b7bb-4fa1-bbb1-77734eed3bd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.664045] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102519, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.684678] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] c7739dfd-88d6-46ac-b3a0-e7dfe9b00332/c7739dfd-88d6-46ac-b3a0-e7dfe9b00332.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.687790] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62ccf590-5eab-42aa-98ca-576d6f7eec23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.702827] env[65107]: DEBUG nova.network.neutron [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Successfully updated port: 454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 723.711417] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 723.711417] env[65107]: value = "task-5102520" [ 723.711417] env[65107]: _type = "Task" [ 723.711417] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.727559] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.764294] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 723.764673] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-801d931d-96a3-42c9-8479-be1465a8d895 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.773305] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 723.773305] env[65107]: value = "task-5102521" [ 723.773305] env[65107]: _type = "Task" [ 723.773305] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.783082] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.796305] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080681} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.796305] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.796305] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe19a685-db51-4633-b3b8-a1bd9a1239db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.820043] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0/c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.821473] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.822101] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.822453] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.835960] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-080bc180-ba1c-41fe-a7b1-d10d8ba64a43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.858389] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527239d0-85e4-0b1f-b6d2-35298a1d5fb6, 'name': SearchDatastore_Task, 'duration_secs': 0.011248} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.859916] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.859916] env[65107]: value = "task-5102522" [ 723.859916] env[65107]: _type = "Task" [ 723.859916] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.861224] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b222049-1b7d-41d3-92df-7858dc278e85 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.875443] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102522, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.875846] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.875846] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fa9ff1-d714-7d9a-714c-e04f6610392c" [ 723.875846] env[65107]: _type = "Task" [ 723.875846] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.889087] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fa9ff1-d714-7d9a-714c-e04f6610392c, 'name': SearchDatastore_Task, 'duration_secs': 0.010984} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.892964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.892964] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 90706f49-af87-4038-ae76-02a95e4a12e6/90706f49-af87-4038-ae76-02a95e4a12e6.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 723.892964] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dda58048-e812-4c9a-b855-5eb63704b4f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.903783] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 723.903783] env[65107]: value = "task-5102523" [ 723.903783] env[65107]: _type = "Task" [ 723.903783] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.914423] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.053933] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6063d581-918c-4120-936d-b7dc1f4dbe05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.062510] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a5a5fc-8468-428c-a76a-10a37d039588 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.097175] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88dccc54-3e9b-45a3-87df-d9b0685f09d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.107386] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993f2b1d-be73-48ec-8c2f-3d885ecf8c5b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.122305] env[65107]: DEBUG nova.compute.provider_tree [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.157667] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102519, 'name': Rename_Task, 'duration_secs': 0.16639} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.157948] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 724.158217] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-514e83d8-e2cc-44e0-a73f-b3f47cc6f7e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.166048] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 724.166048] env[65107]: value = "task-5102524" [ 724.166048] env[65107]: _type = "Task" [ 724.166048] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.175322] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102524, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.206304] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "refresh_cache-854f7a3f-9b7a-4506-8310-94533f0e4cfc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.208034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquired lock "refresh_cache-854f7a3f-9b7a-4506-8310-94533f0e4cfc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.208034] env[65107]: DEBUG nova.network.neutron [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 724.223313] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102520, 'name': ReconfigVM_Task, 'duration_secs': 0.309443} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.223644] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Reconfigured VM instance instance-00000020 to attach disk [datastore1] c7739dfd-88d6-46ac-b3a0-e7dfe9b00332/c7739dfd-88d6-46ac-b3a0-e7dfe9b00332.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.225883] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70a80f35-c1ef-480c-8727-1a17bfe69b98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.234682] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 724.234682] env[65107]: value = "task-5102525" [ 724.234682] env[65107]: _type = "Task" [ 724.234682] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.245339] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102525, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.284834] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.344673] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Updated VIF entry in instance network info cache for port abf46aae-348e-4b4b-acfe-35830ba162e1. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 724.345666] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Updating instance_info_cache with network_info: [{"id": "abf46aae-348e-4b4b-acfe-35830ba162e1", "address": "fa:16:3e:af:30:b1", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabf46aae-34", "ovs_interfaceid": "abf46aae-348e-4b4b-acfe-35830ba162e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.388855] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.416233] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102523, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.657881] env[65107]: DEBUG nova.scheduler.client.report [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 724.658168] env[65107]: DEBUG nova.compute.provider_tree [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 59 to 60 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 724.658346] env[65107]: DEBUG nova.compute.provider_tree [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.677156] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102524, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.711943] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.712363] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.745784] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102525, 'name': Rename_Task, 'duration_secs': 0.297093} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.746107] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 724.746361] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f9f9134-5f81-4f32-803b-655e66f58fb4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.750051] env[65107]: DEBUG nova.network.neutron [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 724.753757] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 724.753757] env[65107]: value = "task-5102526" [ 724.753757] env[65107]: _type = "Task" [ 724.753757] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.765093] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.787477] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.798676] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.799134] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.854682] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Releasing lock "refresh_cache-c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.854984] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Received event network-vif-plugged-53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 724.855373] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquiring lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.855482] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.855653] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.855931] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] No waiting events found dispatching network-vif-plugged-53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 724.856234] env[65107]: WARNING nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Received unexpected event network-vif-plugged-53ef6db1-9d79-43ac-ad0a-c794a9fed04d for instance with vm_state building and task_state spawning. [ 724.856493] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Received event network-changed-53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 724.856723] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Refreshing instance network info cache due to event network-changed-53ef6db1-9d79-43ac-ad0a-c794a9fed04d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 724.857018] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquiring lock "refresh_cache-3948dcdd-e7c1-40b6-a3ec-97741940a0fb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.857206] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquired lock "refresh_cache-3948dcdd-e7c1-40b6-a3ec-97741940a0fb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.857369] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Refreshing network info cache for port 53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 724.876963] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.890638] env[65107]: WARNING neutronclient.v2_0.client [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.891465] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.891865] env[65107]: WARNING openstack [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.917855] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530327} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.918257] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 90706f49-af87-4038-ae76-02a95e4a12e6/90706f49-af87-4038-ae76-02a95e4a12e6.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 724.919893] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.919893] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-759a6a3d-452a-444e-b3d6-b33d48d1f66f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.930558] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 724.930558] env[65107]: value = "task-5102527" [ 724.930558] env[65107]: _type = "Task" [ 724.930558] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.941924] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.964509] env[65107]: DEBUG nova.compute.manager [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Received event network-vif-plugged-52873450-f314-4110-a316-7380569e0afe {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 724.964873] env[65107]: DEBUG oslo_concurrency.lockutils [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Acquiring lock "90706f49-af87-4038-ae76-02a95e4a12e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.965239] env[65107]: DEBUG oslo_concurrency.lockutils [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Lock "90706f49-af87-4038-ae76-02a95e4a12e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.965485] env[65107]: DEBUG oslo_concurrency.lockutils [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Lock "90706f49-af87-4038-ae76-02a95e4a12e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.965942] env[65107]: DEBUG nova.compute.manager [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] No waiting events found dispatching network-vif-plugged-52873450-f314-4110-a316-7380569e0afe {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 724.966157] env[65107]: WARNING nova.compute.manager [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Received unexpected event network-vif-plugged-52873450-f314-4110-a316-7380569e0afe for instance with vm_state building and task_state spawning. [ 724.966287] env[65107]: DEBUG nova.compute.manager [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Received event network-changed-52873450-f314-4110-a316-7380569e0afe {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 724.968116] env[65107]: DEBUG nova.compute.manager [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Refreshing instance network info cache due to event network-changed-52873450-f314-4110-a316-7380569e0afe. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 724.968602] env[65107]: DEBUG oslo_concurrency.lockutils [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Acquiring lock "refresh_cache-90706f49-af87-4038-ae76-02a95e4a12e6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.968793] env[65107]: DEBUG oslo_concurrency.lockutils [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Acquired lock "refresh_cache-90706f49-af87-4038-ae76-02a95e4a12e6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.968994] env[65107]: DEBUG nova.network.neutron [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Refreshing network info cache for port 52873450-f314-4110-a316-7380569e0afe {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 725.043448] env[65107]: DEBUG nova.network.neutron [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Updating instance_info_cache with network_info: [{"id": "454bd66e-49a9-4417-869d-fb1c0f08d472", "address": "fa:16:3e:6a:d2:35", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap454bd66e-49", "ovs_interfaceid": "454bd66e-49a9-4417-869d-fb1c0f08d472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 725.164058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.414s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.164695] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 725.167820] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.492s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.168137] env[65107]: DEBUG nova.objects.instance [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lazy-loading 'resources' on Instance uuid e40dc8ad-31cf-4798-ab97-3ee67c8fbe67 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 725.181066] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102524, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.272094] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102526, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.288884] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102521, 'name': PowerOffVM_Task, 'duration_secs': 1.255964} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.289129] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 725.289356] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.290173] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8819433a-866f-443d-9c6c-1ab77f5c733a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.300379] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 725.300676] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16e10409-90fb-4d07-b712-7d586e6603ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.349191] env[65107]: DEBUG nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 725.349191] env[65107]: DEBUG nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing instance network info cache due to event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 725.349380] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.349543] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.349767] env[65107]: DEBUG nova.network.neutron [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 725.360450] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.361395] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.361497] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.382186] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102522, 'name': ReconfigVM_Task, 'duration_secs': 1.354839} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.384024] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Reconfigured VM instance instance-00000022 to attach disk [datastore1] c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0/c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.384024] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8b06345-40d8-438d-a1d7-d4c46948f465 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.393561] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 725.393561] env[65107]: value = "task-5102529" [ 725.393561] env[65107]: _type = "Task" [ 725.393561] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.405690] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102529, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.413378] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 725.413378] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 725.413378] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleting the datastore file [datastore1] e3510c31-d6be-4e9f-a0a2-a662123861e9 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 725.413378] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46ce9710-b5d5-445b-9454-beb63ae329f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.431024] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 725.431024] env[65107]: value = "task-5102530" [ 725.431024] env[65107]: _type = "Task" [ 725.431024] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.448025] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102530, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.449991] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078496} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.450479] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.451789] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337f43c8-a021-46ec-85e8-02260438ce45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.483960] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] 90706f49-af87-4038-ae76-02a95e4a12e6/90706f49-af87-4038-ae76-02a95e4a12e6.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.489790] env[65107]: WARNING neutronclient.v2_0.client [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.490733] env[65107]: WARNING openstack [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.491178] env[65107]: WARNING openstack [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.498851] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a4d9241-67ff-4425-8bdc-65a3c9877fff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.523553] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 725.523553] env[65107]: value = "task-5102531" [ 725.523553] env[65107]: _type = "Task" [ 725.523553] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.535013] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102531, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.536811] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.537309] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.547367] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Releasing lock "refresh_cache-854f7a3f-9b7a-4506-8310-94533f0e4cfc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.547927] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Instance network_info: |[{"id": "454bd66e-49a9-4417-869d-fb1c0f08d472", "address": "fa:16:3e:6a:d2:35", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap454bd66e-49", "ovs_interfaceid": "454bd66e-49a9-4417-869d-fb1c0f08d472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 725.548667] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:d2:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '454bd66e-49a9-4417-869d-fb1c0f08d472', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.557595] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Creating folder: Project (847cf5be309646b59ed21971ea28f0a2). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 725.557944] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c5aae94-4ff6-47aa-a45d-6e91de6347b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.572108] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Created folder: Project (847cf5be309646b59ed21971ea28f0a2) in parent group-v992574. [ 725.572341] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Creating folder: Instances. Parent ref: group-v992678. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 725.572689] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d9a4d57-74ff-4796-a853-abf60b2006e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.584455] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Created folder: Instances in parent group-v992678. [ 725.584769] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.585045] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 725.585288] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-219d8635-402a-462b-a8af-5ede361013a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.609779] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.609779] env[65107]: value = "task-5102534" [ 725.609779] env[65107]: _type = "Task" [ 725.609779] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.620581] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102534, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.667152] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.667916] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.668600] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.680546] env[65107]: DEBUG nova.compute.utils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 725.682855] env[65107]: WARNING openstack [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.683233] env[65107]: WARNING openstack [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.693513] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 725.693725] env[65107]: DEBUG nova.network.neutron [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 725.694211] env[65107]: WARNING neutronclient.v2_0.client [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.694383] env[65107]: WARNING neutronclient.v2_0.client [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.694998] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.695499] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.702837] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 725.712541] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102524, 'name': PowerOnVM_Task, 'duration_secs': 1.402652} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.712883] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 725.713149] env[65107]: INFO nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Took 10.11 seconds to spawn the instance on the hypervisor. [ 725.713336] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 725.714551] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44a93d1-1e8e-4eaf-9409-ec9d27af7d4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.767945] env[65107]: DEBUG oslo_vmware.api [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102526, 'name': PowerOnVM_Task, 'duration_secs': 0.716211} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.768303] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 725.768553] env[65107]: INFO nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Took 13.10 seconds to spawn the instance on the hypervisor. [ 725.768767] env[65107]: DEBUG nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 725.769733] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b909d38-9c1d-4fcf-b6da-f999add3de32 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.782112] env[65107]: DEBUG nova.policy [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3596ac6ebe3406881fd3a43df23818e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5e0e4c6477c48f8932a11808a42481f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 725.843760] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Updated VIF entry in instance network info cache for port 53ef6db1-9d79-43ac-ad0a-c794a9fed04d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 725.844304] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Updating instance_info_cache with network_info: [{"id": "53ef6db1-9d79-43ac-ad0a-c794a9fed04d", "address": "fa:16:3e:1a:2d:5f", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53ef6db1-9d", "ovs_interfaceid": "53ef6db1-9d79-43ac-ad0a-c794a9fed04d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 725.856859] env[65107]: WARNING neutronclient.v2_0.client [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.857153] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.857309] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.914464] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102529, 'name': Rename_Task, 'duration_secs': 0.26687} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.914464] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 725.914711] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41226820-2930-4baf-ba09-43fcf702d60f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.924225] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 725.924225] env[65107]: value = "task-5102535" [ 725.924225] env[65107]: _type = "Task" [ 725.924225] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.941803] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.948414] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102530, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.041457] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.124948] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102534, 'name': CreateVM_Task, 'duration_secs': 0.511627} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.125186] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 726.125832] env[65107]: WARNING neutronclient.v2_0.client [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.126522] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.127811] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.127811] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 726.127811] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a90b690-1f87-4cde-8362-6d4d82bbacc1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.135859] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 726.135859] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e4022-cb6a-47f6-621b-5408bc3391df" [ 726.135859] env[65107]: _type = "Task" [ 726.135859] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.152300] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e4022-cb6a-47f6-621b-5408bc3391df, 'name': SearchDatastore_Task, 'duration_secs': 0.013629} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.152815] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.153094] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.153342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.153579] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.153843] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.154361] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-124b50fc-8425-456c-a4d6-bfaccba6022d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.168290] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.168635] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 726.169473] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c38afdb-a4b1-4639-b6ca-adfc44186cf9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.183256] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 726.183256] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c3c379-6046-333f-f0e3-42b8ad1908f9" [ 726.183256] env[65107]: _type = "Task" [ 726.183256] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.193096] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c3c379-6046-333f-f0e3-42b8ad1908f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.240205] env[65107]: INFO nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Took 44.31 seconds to build instance. [ 726.271226] env[65107]: WARNING neutronclient.v2_0.client [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.272254] env[65107]: WARNING openstack [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.272687] env[65107]: WARNING openstack [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.295101] env[65107]: INFO nova.compute.manager [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Took 45.46 seconds to build instance. [ 726.347975] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Releasing lock "refresh_cache-3948dcdd-e7c1-40b6-a3ec-97741940a0fb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.348332] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Received event network-vif-plugged-7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 726.348946] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquiring lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.348946] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.348946] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.349112] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] No waiting events found dispatching network-vif-plugged-7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 726.349288] env[65107]: WARNING nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Received unexpected event network-vif-plugged-7b78f034-800f-47c1-9c6f-7e94df8703b1 for instance with vm_state building and task_state spawning. [ 726.349456] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Received event network-changed-7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 726.349609] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Refreshing instance network info cache due to event network-changed-7b78f034-800f-47c1-9c6f-7e94df8703b1. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 726.349804] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquiring lock "refresh_cache-c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.349937] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquired lock "refresh_cache-c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.350117] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Refreshing network info cache for port 7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 726.443165] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102535, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.452148] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102530, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.565178} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.452253] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 726.452833] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 726.452975] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.457700] env[65107]: DEBUG nova.network.neutron [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Successfully created port: b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 726.461507] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b95f412-af9b-4041-8543-a8aee153f7cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.470838] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eaeb8cd-e13a-49cb-9d41-0398ed04bde8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.516115] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550f53d0-1936-4960-b7bd-e406dbcdc097 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.526264] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c750002b-fe59-413e-8848-6a978c114b45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.539276] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102531, 'name': ReconfigVM_Task, 'duration_secs': 0.641227} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.547329] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Reconfigured VM instance instance-00000023 to attach disk [datastore1] 90706f49-af87-4038-ae76-02a95e4a12e6/90706f49-af87-4038-ae76-02a95e4a12e6.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.548252] env[65107]: DEBUG nova.compute.provider_tree [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.549756] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdc8826c-b762-40e0-916f-3224bda9dcf0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.560404] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 726.560404] env[65107]: value = "task-5102536" [ 726.560404] env[65107]: _type = "Task" [ 726.560404] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.570638] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102536, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.690343] env[65107]: DEBUG nova.network.neutron [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Updated VIF entry in instance network info cache for port 52873450-f314-4110-a316-7380569e0afe. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 726.690816] env[65107]: DEBUG nova.network.neutron [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Updating instance_info_cache with network_info: [{"id": "52873450-f314-4110-a316-7380569e0afe", "address": "fa:16:3e:3c:28:f0", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52873450-f3", "ovs_interfaceid": "52873450-f314-4110-a316-7380569e0afe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 726.699691] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c3c379-6046-333f-f0e3-42b8ad1908f9, 'name': SearchDatastore_Task, 'duration_secs': 0.014666} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.701085] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b6d6fed-7c2d-4a1f-b41f-18e3282fb956 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.710162] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 726.710162] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a245c0-49bd-7ecf-353e-2721f248fe62" [ 726.710162] env[65107]: _type = "Task" [ 726.710162] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.715792] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 726.723587] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a245c0-49bd-7ecf-353e-2721f248fe62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.742961] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 726.743219] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.743601] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 726.743668] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.743887] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 726.744156] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 726.744594] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.745650] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 726.745650] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 726.745650] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 726.745650] env[65107]: DEBUG nova.virt.hardware [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 726.746764] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.833s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.748726] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b93843a-6050-4ba3-8e42-7623dbf26618 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.764382] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.766827] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.773796] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e8c448-6f40-4336-9a26-6ce897f00ada {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.797879] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b59619d8-c165-4a2e-8bd8-5747563592e8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.972s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.853022] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.853741] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.855494] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.936260] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102535, 'name': PowerOnVM_Task, 'duration_secs': 0.789261} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.936946] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 726.936946] env[65107]: INFO nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Took 9.12 seconds to spawn the instance on the hypervisor. [ 726.936946] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 726.937771] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f12df91-4008-4416-a179-4196d1a2a436 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.056359] env[65107]: DEBUG nova.scheduler.client.report [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 727.075418] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102536, 'name': Rename_Task, 'duration_secs': 0.177968} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.076580] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 727.076922] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf306b35-07df-47d9-b6ae-e750ad3a65b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.082558] env[65107]: WARNING neutronclient.v2_0.client [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.083255] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.083697] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.093730] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 727.093730] env[65107]: value = "task-5102537" [ 727.093730] env[65107]: _type = "Task" [ 727.093730] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.103247] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102537, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.194432] env[65107]: DEBUG oslo_concurrency.lockutils [req-257c562e-4120-41f6-b1a8-db7d6d1932a8 req-2d0c11b9-73b2-4287-bc11-3a9280c0587b service nova] Releasing lock "refresh_cache-90706f49-af87-4038-ae76-02a95e4a12e6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.209020] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.209020] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.224370] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a245c0-49bd-7ecf-353e-2721f248fe62, 'name': SearchDatastore_Task, 'duration_secs': 0.022319} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.224748] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.225062] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 854f7a3f-9b7a-4506-8310-94533f0e4cfc/854f7a3f-9b7a-4506-8310-94533f0e4cfc.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 727.225384] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07dc1639-5ff2-4d8b-8b9b-73038dd7ee9c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.233346] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 727.233346] env[65107]: value = "task-5102538" [ 727.233346] env[65107]: _type = "Task" [ 727.233346] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.243891] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.245951] env[65107]: DEBUG nova.network.neutron [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updated VIF entry in instance network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 727.249033] env[65107]: DEBUG nova.network.neutron [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 727.257085] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 727.307795] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 727.381549] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.382348] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.382778] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.466340] env[65107]: INFO nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Took 45.49 seconds to build instance. [ 727.507635] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Updated VIF entry in instance network info cache for port 7b78f034-800f-47c1-9c6f-7e94df8703b1. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 727.507635] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Updating instance_info_cache with network_info: [{"id": "7b78f034-800f-47c1-9c6f-7e94df8703b1", "address": "fa:16:3e:39:5f:1a", "network": {"id": "2ac1d31d-6b0c-453c-aea8-b0d8d314e799", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-438989191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c347db6748d044e8842e336a7f25c815", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca50cd14-9e1f-4d74-a066-e5a45ba0ce22", "external-id": "nsx-vlan-transportzone-348", "segmentation_id": 348, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b78f034-80", "ovs_interfaceid": "7b78f034-800f-47c1-9c6f-7e94df8703b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 727.526582] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 727.526884] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 727.527095] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 727.527349] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 727.527510] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 727.527675] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 727.527946] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.528131] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 727.528329] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 727.528501] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 727.528730] env[65107]: DEBUG nova.virt.hardware [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 727.530185] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a714df-c29e-42c9-9ba3-828e59399ac1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.540358] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e08009a-cf61-47fd-babe-0c05366e5b0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.559435] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:7b:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2686e3e0-c528-42bb-9444-050209faa6b1', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.567636] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 727.569063] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.401s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.571334] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.572775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.663s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.578346] env[65107]: INFO nova.compute.claims [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.581686] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42ba26f1-e0b7-4ac6-980e-dde1e97cacec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.609146] env[65107]: INFO nova.scheduler.client.report [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Deleted allocations for instance e40dc8ad-31cf-4798-ab97-3ee67c8fbe67 [ 727.617511] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.617511] env[65107]: value = "task-5102539" [ 727.617511] env[65107]: _type = "Task" [ 727.617511] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.617511] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102537, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.633843] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102539, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.747460] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102538, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.753145] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.753460] env[65107]: DEBUG nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Received event network-vif-plugged-454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 727.753754] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Acquiring lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.754037] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.754258] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.754463] env[65107]: DEBUG nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] No waiting events found dispatching network-vif-plugged-454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 727.754710] env[65107]: WARNING nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Received unexpected event network-vif-plugged-454bd66e-49a9-4417-869d-fb1c0f08d472 for instance with vm_state building and task_state spawning. [ 727.754898] env[65107]: DEBUG nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Received event network-changed-454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 727.755143] env[65107]: DEBUG nova.compute.manager [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Refreshing instance network info cache due to event network-changed-454bd66e-49a9-4417-869d-fb1c0f08d472. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 727.755371] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Acquiring lock "refresh_cache-854f7a3f-9b7a-4506-8310-94533f0e4cfc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.755604] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Acquired lock "refresh_cache-854f7a3f-9b7a-4506-8310-94533f0e4cfc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.755730] env[65107]: DEBUG nova.network.neutron [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Refreshing network info cache for port 454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 727.772089] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.772337] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.787725] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.839302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.969614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "f7446d76-45c2-4e8b-981d-d37c230cf125" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.970372] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.970696] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "f7446d76-45c2-4e8b-981d-d37c230cf125-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.970940] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.972064] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.975552] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.013s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.977918] env[65107]: INFO nova.compute.manager [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Terminating instance [ 728.017604] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Releasing lock "refresh_cache-c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.017604] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 728.017604] env[65107]: DEBUG nova.compute.manager [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing instance network info cache due to event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 728.017604] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.017604] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.018098] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 728.111758] env[65107]: DEBUG oslo_vmware.api [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102537, 'name': PowerOnVM_Task, 'duration_secs': 0.653118} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.112140] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 728.112356] env[65107]: INFO nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Took 8.09 seconds to spawn the instance on the hypervisor. [ 728.112540] env[65107]: DEBUG nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 728.114371] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9195ba-ffc2-4123-b9a7-f32163e81a38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.132022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d069a26-6dd0-4166-82b9-81d4f538c1a3 tempest-ServerMetadataNegativeTestJSON-921490363 tempest-ServerMetadataNegativeTestJSON-921490363-project-member] Lock "e40dc8ad-31cf-4798-ab97-3ee67c8fbe67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.125s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.139252] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102539, 'name': CreateVM_Task, 'duration_secs': 0.50049} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.139749] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 728.141865] env[65107]: WARNING neutronclient.v2_0.client [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 728.141865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.141865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.141865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 728.141865] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dad2e0a-ff24-498e-b227-6ad9b7b72c58 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.147594] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 728.147594] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523c2f4e-e5a1-8eb7-ca84-6e7004ba8d9b" [ 728.147594] env[65107]: _type = "Task" [ 728.147594] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.168483] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523c2f4e-e5a1-8eb7-ca84-6e7004ba8d9b, 'name': SearchDatastore_Task, 'duration_secs': 0.012407} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.168483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.168483] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.168483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.168759] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.168759] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.168759] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51401260-5730-4eac-b137-cf1a23c34c0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.179826] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.180062] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 728.180831] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa4e5cfd-5eb3-48b9-914d-a3075cc0c54c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.189579] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 728.189579] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b0d48-519b-d9d8-f52b-d5fc509094cf" [ 728.189579] env[65107]: _type = "Task" [ 728.189579] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.200641] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b0d48-519b-d9d8-f52b-d5fc509094cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.239785] env[65107]: DEBUG nova.network.neutron [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Successfully updated port: b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 728.248427] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65605} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.248691] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 854f7a3f-9b7a-4506-8310-94533f0e4cfc/854f7a3f-9b7a-4506-8310-94533f0e4cfc.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 728.248912] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.249177] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1dbf4eba-e6ce-4e1c-b60e-010808c7e0fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.259471] env[65107]: WARNING neutronclient.v2_0.client [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.261706] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.262748] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.272399] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 728.272399] env[65107]: value = "task-5102540" [ 728.272399] env[65107]: _type = "Task" [ 728.272399] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.285510] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.455205] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.458841] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.480512] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 728.486942] env[65107]: DEBUG nova.compute.manager [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 728.486942] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.486942] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53584760-60d3-46b2-b8cf-c16dcf614808 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.494630] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 728.495238] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc85d276-5754-4afe-9677-e6d73ff8dc97 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.503584] env[65107]: DEBUG oslo_vmware.api [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 728.503584] env[65107]: value = "task-5102541" [ 728.503584] env[65107]: _type = "Task" [ 728.503584] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.515057] env[65107]: DEBUG oslo_vmware.api [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102541, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.522941] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.522941] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.522941] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.641196] env[65107]: INFO nova.compute.manager [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Took 43.12 seconds to build instance. [ 728.664032] env[65107]: WARNING neutronclient.v2_0.client [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.664032] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.664032] env[65107]: WARNING openstack [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.704636] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b0d48-519b-d9d8-f52b-d5fc509094cf, 'name': SearchDatastore_Task, 'duration_secs': 0.014116} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.705933] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bf5f694-d976-46cf-b13a-937c9af89fd6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.718219] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 728.718219] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0403d-36cf-f0a6-0200-0eb5a4ab101b" [ 728.718219] env[65107]: _type = "Task" [ 728.718219] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.732782] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0403d-36cf-f0a6-0200-0eb5a4ab101b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.744914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "refresh_cache-d40aec5a-d898-42f3-a121-503cf043f72b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.745085] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquired lock "refresh_cache-d40aec5a-d898-42f3-a121-503cf043f72b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.745577] env[65107]: DEBUG nova.network.neutron [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 728.786114] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080446} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.789447] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 728.791103] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619d7f3b-9f44-466a-9e21-aa9bd8eff9be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.820354] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 854f7a3f-9b7a-4506-8310-94533f0e4cfc/854f7a3f-9b7a-4506-8310-94533f0e4cfc.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.823780] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b945c4c8-209a-4418-a5c7-1233a21c79c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.848471] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 728.848471] env[65107]: value = "task-5102542" [ 728.848471] env[65107]: _type = "Task" [ 728.848471] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.859823] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102542, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.865959] env[65107]: DEBUG nova.compute.manager [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 728.866175] env[65107]: DEBUG nova.compute.manager [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing instance network info cache due to event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 728.868907] env[65107]: DEBUG oslo_concurrency.lockutils [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.944617] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.945975] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.969541] env[65107]: DEBUG nova.network.neutron [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Updated VIF entry in instance network info cache for port 454bd66e-49a9-4417-869d-fb1c0f08d472. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 728.970163] env[65107]: DEBUG nova.network.neutron [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Updating instance_info_cache with network_info: [{"id": "454bd66e-49a9-4417-869d-fb1c0f08d472", "address": "fa:16:3e:6a:d2:35", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap454bd66e-49", "ovs_interfaceid": "454bd66e-49a9-4417-869d-fb1c0f08d472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.011504] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.022085] env[65107]: DEBUG oslo_vmware.api [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102541, 'name': PowerOffVM_Task, 'duration_secs': 0.418607} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.022085] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 729.022085] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 729.022702] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0cd48ca-cdf1-409b-83c2-2cef59795c9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.109535] env[65107]: WARNING neutronclient.v2_0.client [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.110156] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.110486] env[65107]: WARNING openstack [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.124401] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 729.124630] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 729.124823] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Deleting the datastore file [datastore2] f7446d76-45c2-4e8b-981d-d37c230cf125 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 729.125155] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f2624bd-372c-4049-ac99-9ec6c552f09d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.133838] env[65107]: DEBUG oslo_vmware.api [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for the task: (returnval){ [ 729.133838] env[65107]: value = "task-5102544" [ 729.133838] env[65107]: _type = "Task" [ 729.133838] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.149942] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8d85aa8b-1409-4e36-8d0d-6df5f32907a1 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "90706f49-af87-4038-ae76-02a95e4a12e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.146s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.150315] env[65107]: DEBUG oslo_vmware.api [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102544, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.241173] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0403d-36cf-f0a6-0200-0eb5a4ab101b, 'name': SearchDatastore_Task, 'duration_secs': 0.048316} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.241582] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.242207] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.242517] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b98949a-4ccd-4d2d-ad3a-22eb0f1e5c12 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.249705] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.250283] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.259674] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 729.259674] env[65107]: value = "task-5102545" [ 729.259674] env[65107]: _type = "Task" [ 729.259674] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.276327] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.349400] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updated VIF entry in instance network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 729.349780] env[65107]: DEBUG nova.network.neutron [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.362374] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102542, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.418082] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a164952-ab3f-4e87-8329-a429cff49a43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.429689] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286557fe-5cdd-4054-9c54-f1d94dfb1d56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.467158] env[65107]: DEBUG nova.network.neutron [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 729.470051] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b0343e-5fd5-40d5-b834-11c39f3c57a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.475147] env[65107]: DEBUG oslo_concurrency.lockutils [req-af28a0b0-41e0-494f-9cbc-e04729bd6caf req-51350b20-3c64-4f40-8168-f8fce4a6f681 service nova] Releasing lock "refresh_cache-854f7a3f-9b7a-4506-8310-94533f0e4cfc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.479069] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa8f65a-90dc-4e72-9be7-1e5546197630 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.495123] env[65107]: DEBUG nova.compute.provider_tree [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.498092] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.498578] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.510087] env[65107]: DEBUG nova.scheduler.client.report [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 729.653035] env[65107]: DEBUG oslo_vmware.api [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Task: {'id': task-5102544, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253356} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.653035] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 729.653035] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 729.653035] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.653035] env[65107]: INFO nova.compute.manager [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Took 1.17 seconds to destroy the instance on the hypervisor. [ 729.653313] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 729.653313] env[65107]: DEBUG nova.compute.manager [-] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 729.653313] env[65107]: DEBUG nova.network.neutron [-] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 729.653313] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.653313] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.653313] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.660831] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 729.680372] env[65107]: WARNING neutronclient.v2_0.client [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.681104] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.682297] env[65107]: WARNING openstack [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.774828] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102545, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.790019] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.856525] env[65107]: DEBUG oslo_concurrency.lockutils [req-e327d1c7-3a02-458c-a2c8-4db332286682 req-1f223cab-7830-4c74-b8fb-6318270cf031 service nova] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.857479] env[65107]: DEBUG oslo_concurrency.lockutils [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.857752] env[65107]: DEBUG nova.network.neutron [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 729.865191] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102542, 'name': ReconfigVM_Task, 'duration_secs': 0.556733} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.867702] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 854f7a3f-9b7a-4506-8310-94533f0e4cfc/854f7a3f-9b7a-4506-8310-94533f0e4cfc.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.867702] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6aaaa03b-992e-499f-9151-6bdb1b31ed6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.876768] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 729.876768] env[65107]: value = "task-5102546" [ 729.876768] env[65107]: _type = "Task" [ 729.876768] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.889149] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102546, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.015991] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.017019] env[65107]: DEBUG nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 730.020613] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.260s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.020812] env[65107]: DEBUG nova.objects.instance [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lazy-loading 'resources' on Instance uuid a2b72bac-6806-4b7b-b972-32aea52f7c82 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 730.228859] env[65107]: DEBUG nova.network.neutron [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Updating instance_info_cache with network_info: [{"id": "b3b4ec0d-49c0-4c0a-845c-157dd80604fb", "address": "fa:16:3e:93:f8:7e", "network": {"id": "e3e2350e-cc95-4eea-b11a-24dd93d08a32", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-598474907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5e0e4c6477c48f8932a11808a42481f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b4ec0d-49", "ovs_interfaceid": "b3b4ec0d-49c0-4c0a-845c-157dd80604fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.274705] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607771} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.275161] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 730.275486] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 730.275774] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dec4c296-cc2b-4632-b5e6-7cbe353ff6c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.279100] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.283428] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 730.283428] env[65107]: value = "task-5102547" [ 730.283428] env[65107]: _type = "Task" [ 730.283428] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.293589] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.360757] env[65107]: WARNING neutronclient.v2_0.client [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.361855] env[65107]: WARNING openstack [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.362027] env[65107]: WARNING openstack [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.394560] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102546, 'name': Rename_Task, 'duration_secs': 0.220636} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.394560] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 730.394971] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c64b3cef-ba52-4f84-bba0-8896e3712860 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.405191] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 730.405191] env[65107]: value = "task-5102548" [ 730.405191] env[65107]: _type = "Task" [ 730.405191] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.418560] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.524255] env[65107]: DEBUG nova.compute.utils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 730.533560] env[65107]: DEBUG nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 730.687331] env[65107]: WARNING openstack [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.688597] env[65107]: WARNING openstack [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.696287] env[65107]: DEBUG nova.network.neutron [-] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.731283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Releasing lock "refresh_cache-d40aec5a-d898-42f3-a121-503cf043f72b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.731634] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Instance network_info: |[{"id": "b3b4ec0d-49c0-4c0a-845c-157dd80604fb", "address": "fa:16:3e:93:f8:7e", "network": {"id": "e3e2350e-cc95-4eea-b11a-24dd93d08a32", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-598474907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5e0e4c6477c48f8932a11808a42481f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b4ec0d-49", "ovs_interfaceid": "b3b4ec0d-49c0-4c0a-845c-157dd80604fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 730.733262] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:f8:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3b4ec0d-49c0-4c0a-845c-157dd80604fb', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 730.740824] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Creating folder: Project (d5e0e4c6477c48f8932a11808a42481f). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 730.741299] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb391d52-02d8-469a-9434-d51721d59d77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.756622] env[65107]: WARNING neutronclient.v2_0.client [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.756622] env[65107]: WARNING openstack [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.756918] env[65107]: WARNING openstack [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.768882] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Created folder: Project (d5e0e4c6477c48f8932a11808a42481f) in parent group-v992574. [ 730.769098] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Creating folder: Instances. Parent ref: group-v992682. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 730.769530] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7562bcf5-34a1-4aac-a81c-4e20fe8e3b77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.781230] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Created folder: Instances in parent group-v992682. [ 730.781558] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 730.782042] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 730.782042] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc708e5b-4774-48f6-8c58-ec5e92cff55e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.810662] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 730.810662] env[65107]: value = "task-5102551" [ 730.810662] env[65107]: _type = "Task" [ 730.810662] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.823548] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.210054} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.823761] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.826131] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe77560-6453-4deb-9706-41c3eaea5229 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.834984] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102551, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.864029] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.868634] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85dd3bee-ec4a-4bf3-86c2-dab237fbfdb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.888636] env[65107]: DEBUG nova.network.neutron [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updated VIF entry in instance network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 730.890247] env[65107]: DEBUG nova.network.neutron [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.897549] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 730.897549] env[65107]: value = "task-5102552" [ 730.897549] env[65107]: _type = "Task" [ 730.897549] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.910298] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.924981] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102548, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.034492] env[65107]: DEBUG nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 731.201995] env[65107]: INFO nova.compute.manager [-] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Took 1.55 seconds to deallocate network for instance. [ 731.322117] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fb8afa-885a-439c-b90c-f47bb65970ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.328938] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102551, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.335864] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4da59f3-0a47-4b86-a46d-df64d273f4f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.369254] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27f43cd-fffb-45cc-aa92-c14a26b968ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.377577] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e348f98a-ce5d-4512-baaf-03a677cef37a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.393927] env[65107]: DEBUG oslo_concurrency.lockutils [req-27a891dc-47d4-4f1e-bea7-ce89365c3c4d req-5d406aa2-9dbc-4ef7-872e-141a953467f0 service nova] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.394665] env[65107]: DEBUG nova.compute.provider_tree [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.407266] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102552, 'name': ReconfigVM_Task, 'duration_secs': 0.376658} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.408248] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Reconfigured VM instance instance-00000002 to attach disk [datastore2] e3510c31-d6be-4e9f-a0a2-a662123861e9/e3510c31-d6be-4e9f-a0a2-a662123861e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.408910] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-057368f6-3f44-4ec9-8bc0-82e28a2d3010 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.422229] env[65107]: DEBUG oslo_vmware.api [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102548, 'name': PowerOnVM_Task, 'duration_secs': 0.912875} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.423355] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 731.423565] env[65107]: INFO nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Took 8.64 seconds to spawn the instance on the hypervisor. [ 731.423742] env[65107]: DEBUG nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 731.424090] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 731.424090] env[65107]: value = "task-5102553" [ 731.424090] env[65107]: _type = "Task" [ 731.424090] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.424819] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1424393-75b9-4e2c-bd16-0c706c1bccc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.437139] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102553, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.563817] env[65107]: DEBUG nova.compute.manager [req-94ebe15d-a068-4815-b4ef-78287037f1fd req-fa115d8e-92d6-4116-8775-ded12fb8a547 service nova] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Received event network-vif-deleted-8a1e7b3c-e981-43b6-84a0-1adcd2b2a044 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 731.630259] env[65107]: DEBUG oslo_concurrency.lockutils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.630384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.713653] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.736551] env[65107]: DEBUG nova.compute.manager [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Received event network-vif-plugged-b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 731.737075] env[65107]: DEBUG oslo_concurrency.lockutils [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Acquiring lock "d40aec5a-d898-42f3-a121-503cf043f72b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.737645] env[65107]: DEBUG oslo_concurrency.lockutils [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Lock "d40aec5a-d898-42f3-a121-503cf043f72b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.737830] env[65107]: DEBUG oslo_concurrency.lockutils [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Lock "d40aec5a-d898-42f3-a121-503cf043f72b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.738050] env[65107]: DEBUG nova.compute.manager [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] No waiting events found dispatching network-vif-plugged-b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 731.738220] env[65107]: WARNING nova.compute.manager [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Received unexpected event network-vif-plugged-b3b4ec0d-49c0-4c0a-845c-157dd80604fb for instance with vm_state building and task_state spawning. [ 731.738390] env[65107]: DEBUG nova.compute.manager [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Received event network-changed-b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 731.738530] env[65107]: DEBUG nova.compute.manager [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Refreshing instance network info cache due to event network-changed-b3b4ec0d-49c0-4c0a-845c-157dd80604fb. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 731.738765] env[65107]: DEBUG oslo_concurrency.lockutils [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Acquiring lock "refresh_cache-d40aec5a-d898-42f3-a121-503cf043f72b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.738848] env[65107]: DEBUG oslo_concurrency.lockutils [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Acquired lock "refresh_cache-d40aec5a-d898-42f3-a121-503cf043f72b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 731.739008] env[65107]: DEBUG nova.network.neutron [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Refreshing network info cache for port b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 731.830163] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102551, 'name': CreateVM_Task, 'duration_secs': 0.527497} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.830163] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 731.830740] env[65107]: WARNING neutronclient.v2_0.client [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 731.831129] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.831286] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 731.832792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 731.833384] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-469cecd2-ee85-4fb8-8ed4-81fabc0600d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.839466] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 731.839466] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523856e3-5ea2-2ac5-fe1e-dab7c811b079" [ 731.839466] env[65107]: _type = "Task" [ 731.839466] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.848833] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523856e3-5ea2-2ac5-fe1e-dab7c811b079, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.900045] env[65107]: DEBUG nova.scheduler.client.report [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.940014] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102553, 'name': Rename_Task, 'duration_secs': 0.154196} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.940729] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 731.941088] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6518f6b-ed64-470a-b4ea-cf99ed3acd43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.951461] env[65107]: INFO nova.compute.manager [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Took 45.98 seconds to build instance. [ 731.954628] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 731.954628] env[65107]: value = "task-5102554" [ 731.954628] env[65107]: _type = "Task" [ 731.954628] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.965059] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.049248] env[65107]: DEBUG nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 732.088728] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 732.089198] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 732.089808] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 732.090921] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 732.091215] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 732.091492] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 732.092103] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.092282] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 732.092584] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 732.092876] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 732.093283] env[65107]: DEBUG nova.virt.hardware [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 732.094715] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e48920-2e52-4583-8809-54a88098f59e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.106912] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574d5dfc-d557-4500-a2b4-395e1021a570 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.129227] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.135140] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Creating folder: Project (d3a3b038495a4682aa19fa3f94c0dec9). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.136391] env[65107]: DEBUG nova.compute.utils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 732.137746] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfae30bb-ce06-4951-ad54-a4e2f48299c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.152449] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Created folder: Project (d3a3b038495a4682aa19fa3f94c0dec9) in parent group-v992574. [ 732.152826] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Creating folder: Instances. Parent ref: group-v992685. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.153286] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25c35f54-6b35-4183-8456-3338c2a28c39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.165380] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Created folder: Instances in parent group-v992685. [ 732.165741] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 732.167163] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 732.167163] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0fa2288a-588d-4f1a-be40-ba9bc3c77bd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.187653] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.187653] env[65107]: value = "task-5102557" [ 732.187653] env[65107]: _type = "Task" [ 732.187653] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.199065] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102557, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.243161] env[65107]: WARNING neutronclient.v2_0.client [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.246433] env[65107]: WARNING openstack [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.246433] env[65107]: WARNING openstack [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.352601] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523856e3-5ea2-2ac5-fe1e-dab7c811b079, 'name': SearchDatastore_Task, 'duration_secs': 0.030305} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.353721] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 732.353721] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.353721] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.353721] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 732.353890] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.353994] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c301cdca-d93e-4f6e-b4e2-b2b6b3372fe4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.364420] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.364727] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 732.365756] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-956ec430-08b0-4083-a997-f2266c125b65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.373155] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 732.373155] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ccbad1-7430-5a39-61e8-52117a51a74d" [ 732.373155] env[65107]: _type = "Task" [ 732.373155] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.383161] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ccbad1-7430-5a39-61e8-52117a51a74d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.406045] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.384s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.408679] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.108s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.410124] env[65107]: INFO nova.compute.claims [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.444389] env[65107]: INFO nova.scheduler.client.report [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Deleted allocations for instance a2b72bac-6806-4b7b-b972-32aea52f7c82 [ 732.453666] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c26afe59-999a-4856-9bac-1214df9ff3d9 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.498s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.470898] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102554, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.641489] env[65107]: DEBUG oslo_concurrency.lockutils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.649366] env[65107]: WARNING openstack [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.649366] env[65107]: WARNING openstack [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.703242] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102557, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.728320] env[65107]: WARNING neutronclient.v2_0.client [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.729013] env[65107]: WARNING openstack [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.729423] env[65107]: WARNING openstack [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.893692] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ccbad1-7430-5a39-61e8-52117a51a74d, 'name': SearchDatastore_Task, 'duration_secs': 0.014808} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.895554] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ab481a6-8600-4415-a39c-50b7c0f74304 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.904359] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 732.904359] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298dc7e-8619-a72b-5a5f-d8c6410a5259" [ 732.904359] env[65107]: _type = "Task" [ 732.904359] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.917469] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298dc7e-8619-a72b-5a5f-d8c6410a5259, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.959029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e4865a87-2935-4cad-8829-0a4791de739e tempest-ServerDiagnosticsV248Test-644351400 tempest-ServerDiagnosticsV248Test-644351400-project-member] Lock "a2b72bac-6806-4b7b-b972-32aea52f7c82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.492s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.962199] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 732.975881] env[65107]: DEBUG oslo_vmware.api [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102554, 'name': PowerOnVM_Task, 'duration_secs': 0.610521} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.976717] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 732.977070] env[65107]: DEBUG nova.compute.manager [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 732.978234] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19df1fe2-fa90-4db2-a8e9-3e20ece7db32 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.055425] env[65107]: DEBUG nova.network.neutron [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Updated VIF entry in instance network info cache for port b3b4ec0d-49c0-4c0a-845c-157dd80604fb. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 733.055425] env[65107]: DEBUG nova.network.neutron [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Updating instance_info_cache with network_info: [{"id": "b3b4ec0d-49c0-4c0a-845c-157dd80604fb", "address": "fa:16:3e:93:f8:7e", "network": {"id": "e3e2350e-cc95-4eea-b11a-24dd93d08a32", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-598474907-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5e0e4c6477c48f8932a11808a42481f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b4ec0d-49", "ovs_interfaceid": "b3b4ec0d-49c0-4c0a-845c-157dd80604fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 733.163293] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.163391] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.163630] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.163858] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.165408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.171319] env[65107]: INFO nova.compute.manager [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Terminating instance [ 733.205743] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102557, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.230332] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "65d249a9-f79c-46c0-8630-169937cbcaf7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.230754] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "65d249a9-f79c-46c0-8630-169937cbcaf7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.358872] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.359290] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.359643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.359970] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.360261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.364424] env[65107]: INFO nova.compute.manager [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Terminating instance [ 733.425580] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298dc7e-8619-a72b-5a5f-d8c6410a5259, 'name': SearchDatastore_Task, 'duration_secs': 0.012483} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.428608] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 733.428608] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d40aec5a-d898-42f3-a121-503cf043f72b/d40aec5a-d898-42f3-a121-503cf043f72b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 733.428608] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dcca96fc-89ff-422c-bfc0-acaad0de3c85 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.437949] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 733.437949] env[65107]: value = "task-5102558" [ 733.437949] env[65107]: _type = "Task" [ 733.437949] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.454522] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.487301] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.509743] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.558426] env[65107]: DEBUG oslo_concurrency.lockutils [req-fea8707a-c9e0-4118-9124-d29a2c9ba9e3 req-49950d54-4dfa-45c4-bb16-097b798d8171 service nova] Releasing lock "refresh_cache-d40aec5a-d898-42f3-a121-503cf043f72b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 733.680077] env[65107]: DEBUG nova.compute.manager [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 733.683120] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.683120] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccab2160-4999-4e59-8cd5-89fdeede0205 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.709883] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102557, 'name': CreateVM_Task, 'duration_secs': 1.145796} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.711206] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 733.713933] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 733.713933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.713933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 733.713933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 733.713933] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2474b6c-2447-4f09-956b-6d983dc16209 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.721579] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfc89213-25b4-4c58-bd71-79e89988c982 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.727689] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 733.727689] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525acb66-4b6d-1ea7-985d-e4e806ece2ca" [ 733.727689] env[65107]: _type = "Task" [ 733.727689] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.733145] env[65107]: DEBUG oslo_vmware.api [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 733.733145] env[65107]: value = "task-5102559" [ 733.733145] env[65107]: _type = "Task" [ 733.733145] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.742033] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525acb66-4b6d-1ea7-985d-e4e806ece2ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.746523] env[65107]: DEBUG oslo_concurrency.lockutils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.746767] env[65107]: DEBUG oslo_concurrency.lockutils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.747899] env[65107]: INFO nova.compute.manager [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Attaching volume 9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc to /dev/sdb [ 733.753551] env[65107]: DEBUG oslo_vmware.api [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102559, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.803714] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968d30b5-e599-42cb-ab36-7957b344b36a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.815939] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcff23f-e67a-4097-9d02-724c413393da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.838742] env[65107]: DEBUG nova.virt.block_device [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Updating existing volume attachment record: e2fd00a3-1e0d-4608-8ee4-89bb822ffd01 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 733.871252] env[65107]: DEBUG nova.compute.manager [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 733.871252] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.873016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df8ff6b-a60b-482b-b3a5-4fa48ef793a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.883615] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 733.889752] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23b3a8f9-ad35-493e-a1c6-9fccc9240ff3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.900722] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 733.900722] env[65107]: value = "task-5102560" [ 733.900722] env[65107]: _type = "Task" [ 733.900722] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.920192] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.957977] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102558, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.255340] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525acb66-4b6d-1ea7-985d-e4e806ece2ca, 'name': SearchDatastore_Task, 'duration_secs': 0.014729} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.255340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 734.255340] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.255340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.255589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.255589] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.259073] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5aba423b-72b7-4efa-a872-0e40f32edee9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.261160] env[65107]: DEBUG oslo_vmware.api [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102559, 'name': PowerOffVM_Task, 'duration_secs': 0.283126} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.261426] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.261590] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 734.262275] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7be50fc3-384d-4c93-8097-25af521a7c69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.282753] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.282753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 734.285191] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-597032f2-05db-4070-8375-8600ab3639aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.297161] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 734.297161] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ca8359-3cee-45ad-6465-44d92c66bb9d" [ 734.297161] env[65107]: _type = "Task" [ 734.297161] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.312094] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ca8359-3cee-45ad-6465-44d92c66bb9d, 'name': SearchDatastore_Task, 'duration_secs': 0.012055} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.312094] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6d9de04-a494-437c-a1f5-9902d4c1aa44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.318686] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 734.318686] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be9cf0-9a4c-5b1b-81e0-f356f8093b33" [ 734.318686] env[65107]: _type = "Task" [ 734.318686] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.340297] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be9cf0-9a4c-5b1b-81e0-f356f8093b33, 'name': SearchDatastore_Task, 'duration_secs': 0.013202} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.340747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 734.343176] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d/c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 734.343622] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cf94160-0f26-4537-b5db-5e7be9b3d8be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.351965] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 734.351965] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 734.351965] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Deleting the datastore file [datastore1] 705b91ea-ccd2-41ce-a14f-2ea5b70f80df {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 734.351965] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4eb13a09-9103-4e63-b055-88a57a04295e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.358237] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 734.358237] env[65107]: value = "task-5102565" [ 734.358237] env[65107]: _type = "Task" [ 734.358237] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.360033] env[65107]: DEBUG oslo_vmware.api [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for the task: (returnval){ [ 734.360033] env[65107]: value = "task-5102566" [ 734.360033] env[65107]: _type = "Task" [ 734.360033] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.380089] env[65107]: DEBUG oslo_vmware.api [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.384582] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.393178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe8480d-02ca-4880-b7a3-8f8e069d325b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.412847] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9431f3-af63-44fe-a2ec-5c685dfc660d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.423666] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102560, 'name': PowerOffVM_Task, 'duration_secs': 0.298381} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.452248] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.452371] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 734.453588] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43d0d072-e3d6-43e1-bf97-647f3d9dc550 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.458868] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30004ce2-a598-403b-bd8f-7a655251f66b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.470870] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727157} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.472496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e14972-cac6-4c51-a4d7-12fe8c675e38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.477567] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d40aec5a-d898-42f3-a121-503cf043f72b/d40aec5a-d898-42f3-a121-503cf043f72b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 734.478492] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.478492] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-918a6831-dcf6-479f-8847-402ab3899979 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.498299] env[65107]: DEBUG nova.compute.provider_tree [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.504391] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 734.504391] env[65107]: value = "task-5102568" [ 734.504391] env[65107]: _type = "Task" [ 734.504391] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.517176] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.546359] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 734.546359] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 734.546766] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleting the datastore file [datastore1] 3948dcdd-e7c1-40b6-a3ec-97741940a0fb {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 734.546970] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bf24897-6753-4d51-b4b2-d424afe69149 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.558384] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 734.558384] env[65107]: value = "task-5102569" [ 734.558384] env[65107]: _type = "Task" [ 734.558384] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.573913] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.577859] env[65107]: DEBUG nova.compute.manager [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 734.578249] env[65107]: DEBUG nova.compute.manager [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing instance network info cache due to event network-changed-41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 734.578626] env[65107]: DEBUG oslo_concurrency.lockutils [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Acquiring lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.578855] env[65107]: DEBUG oslo_concurrency.lockutils [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Acquired lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.579151] env[65107]: DEBUG nova.network.neutron [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Refreshing network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 734.836837] env[65107]: DEBUG nova.compute.manager [None req-95d26eb1-d9ad-42ce-85f9-9693a722a31a tempest-ServerDiagnosticsTest-1095076073 tempest-ServerDiagnosticsTest-1095076073-project-admin] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 734.838800] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a8ea19-762a-403f-a99c-fa4bc823708c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.849882] env[65107]: INFO nova.compute.manager [None req-95d26eb1-d9ad-42ce-85f9-9693a722a31a tempest-ServerDiagnosticsTest-1095076073 tempest-ServerDiagnosticsTest-1095076073-project-admin] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Retrieving diagnostics [ 734.850754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226c95db-ac10-4f83-860c-b0e4dc5d1deb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.904616] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102565, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.905145] env[65107]: DEBUG oslo_vmware.api [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Task: {'id': task-5102566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211123} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.905262] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.905525] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 734.905899] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.905899] env[65107]: INFO nova.compute.manager [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Took 1.23 seconds to destroy the instance on the hypervisor. [ 734.906586] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 734.906586] env[65107]: DEBUG nova.compute.manager [-] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 734.906586] env[65107]: DEBUG nova.network.neutron [-] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 734.907596] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.907596] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.907873] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.003939] env[65107]: DEBUG nova.scheduler.client.report [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 735.019721] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078639} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.019721] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.020025] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207ab866-c3a5-456a-a14b-c73d48e5502a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.050054] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] d40aec5a-d898-42f3-a121-503cf043f72b/d40aec5a-d898-42f3-a121-503cf043f72b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.050889] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2865b17-8a78-47c7-8eb0-708bc6e9c781 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.071238] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.077236] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 735.077236] env[65107]: value = "task-5102570" [ 735.077236] env[65107]: _type = "Task" [ 735.077236] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.081725] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.085454] env[65107]: WARNING neutronclient.v2_0.client [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.086204] env[65107]: WARNING openstack [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.086988] env[65107]: WARNING openstack [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.103347] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.169431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "b20a2c71-514d-488a-8edb-8fb0274a894a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.169431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.169431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "b20a2c71-514d-488a-8edb-8fb0274a894a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.169939] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.169939] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.175184] env[65107]: INFO nova.compute.manager [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Terminating instance [ 735.240959] env[65107]: WARNING openstack [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.240959] env[65107]: WARNING openstack [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.310826] env[65107]: WARNING neutronclient.v2_0.client [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.312232] env[65107]: WARNING openstack [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.312232] env[65107]: WARNING openstack [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.397088] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102565, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561188} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.397841] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d/c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 735.397841] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 735.397965] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8e561f2-7dad-4f68-8617-31abbe248ceb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.406979] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 735.406979] env[65107]: value = "task-5102571" [ 735.406979] env[65107]: _type = "Task" [ 735.406979] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.417530] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.432899] env[65107]: DEBUG nova.network.neutron [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updated VIF entry in instance network info cache for port 41ad83e5-4645-4e91-a223-473eefb69c38. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 735.433414] env[65107]: DEBUG nova.network.neutron [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [{"id": "41ad83e5-4645-4e91-a223-473eefb69c38", "address": "fa:16:3e:27:9f:84", "network": {"id": "658c7491-bf1d-4279-84bf-bd507db11f35", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-963707128-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a4457697290e40e89d4261fe02ff5eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41ad83e5-46", "ovs_interfaceid": "41ad83e5-4645-4e91-a223-473eefb69c38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 735.514404] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.106s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.515038] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 735.517986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.767s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.519526] env[65107]: INFO nova.compute.claims [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.578845] env[65107]: DEBUG oslo_vmware.api [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.559264} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.579202] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 735.579326] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 735.579516] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.579669] env[65107]: INFO nova.compute.manager [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Took 1.71 seconds to destroy the instance on the hypervisor. [ 735.579918] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 735.580403] env[65107]: DEBUG nova.compute.manager [-] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 735.580506] env[65107]: DEBUG nova.network.neutron [-] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 735.580761] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.581304] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.581556] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.598915] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.682027] env[65107]: DEBUG nova.compute.manager [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 735.682027] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.682027] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c0e38a-6f3a-49c1-81ca-00269b18e237 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.692474] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 735.692474] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c513e17-926b-40fa-8f28-1e6d75628a4d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.695957] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.707571] env[65107]: DEBUG oslo_vmware.api [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 735.707571] env[65107]: value = "task-5102572" [ 735.707571] env[65107]: _type = "Task" [ 735.707571] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.718115] env[65107]: DEBUG oslo_vmware.api [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.925049] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.223869} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.925370] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.927396] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b41700-912c-4cc5-908d-e8f43239b8f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.943019] env[65107]: DEBUG oslo_concurrency.lockutils [req-b399943a-6bac-4fbf-870c-6da0a64a76ec req-effcc9c9-1149-4552-b7dc-a9c1a7ba292e service nova] Releasing lock "refresh_cache-b20a2c71-514d-488a-8edb-8fb0274a894a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 735.951869] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d/c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.953630] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0221e992-8b8a-4637-8bd0-0567c5d7785a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.984069] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 735.984069] env[65107]: value = "task-5102573" [ 735.984069] env[65107]: _type = "Task" [ 735.984069] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.994937] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102573, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.026570] env[65107]: DEBUG nova.compute.utils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 736.032201] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 736.032445] env[65107]: DEBUG nova.network.neutron [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 736.032728] env[65107]: WARNING neutronclient.v2_0.client [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.033254] env[65107]: WARNING neutronclient.v2_0.client [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.035135] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.035135] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.103697] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102570, 'name': ReconfigVM_Task, 'duration_secs': 0.976482} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.105024] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Reconfigured VM instance instance-00000025 to attach disk [datastore2] d40aec5a-d898-42f3-a121-503cf043f72b/d40aec5a-d898-42f3-a121-503cf043f72b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.105024] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94e592a0-ae8c-42a0-985b-8990d874b1a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.114148] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 736.114148] env[65107]: value = "task-5102574" [ 736.114148] env[65107]: _type = "Task" [ 736.114148] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.129389] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102574, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.154383] env[65107]: DEBUG nova.network.neutron [-] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 736.160094] env[65107]: DEBUG nova.policy [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66f11f09dd98447bb9851bee92183dcc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2882c2c7949045519a146655bb694a3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 736.228647] env[65107]: DEBUG oslo_vmware.api [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102572, 'name': PowerOffVM_Task, 'duration_secs': 0.313114} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.228647] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 736.228647] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 736.229386] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70eed151-bea2-4fce-a7eb-99d11ee48728 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.314893] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 736.314893] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 736.314893] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Deleting the datastore file [datastore2] b20a2c71-514d-488a-8edb-8fb0274a894a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 736.314893] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-caf33380-4dff-4fd2-9e39-96469d37a47b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.335448] env[65107]: DEBUG oslo_vmware.api [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for the task: (returnval){ [ 736.335448] env[65107]: value = "task-5102576" [ 736.335448] env[65107]: _type = "Task" [ 736.335448] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.351736] env[65107]: DEBUG oslo_vmware.api [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102576, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.473647] env[65107]: DEBUG nova.network.neutron [-] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 736.505143] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102573, 'name': ReconfigVM_Task, 'duration_secs': 0.483853} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.508039] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Reconfigured VM instance instance-00000026 to attach disk [datastore2] c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d/c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.509291] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82decebc-b11a-49c4-83ac-20806e99704d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.519022] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 736.519022] env[65107]: value = "task-5102578" [ 736.519022] env[65107]: _type = "Task" [ 736.519022] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.528417] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102578, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.532741] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 736.629575] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102574, 'name': Rename_Task, 'duration_secs': 0.246665} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.630037] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 736.630582] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8488f131-0ed6-40a0-a72c-18e799d7b52c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.639255] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 736.639255] env[65107]: value = "task-5102579" [ 736.639255] env[65107]: _type = "Task" [ 736.639255] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.651696] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.663919] env[65107]: INFO nova.compute.manager [-] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Took 1.76 seconds to deallocate network for instance. [ 736.732037] env[65107]: DEBUG nova.compute.manager [req-476ca3af-1e8c-47ed-8f77-dfa2a440753a req-8f1fbee7-1876-48e0-988e-87e715f46621 service nova] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Received event network-vif-deleted-53ef6db1-9d79-43ac-ad0a-c794a9fed04d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 736.828076] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.828529] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.830044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.830044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.830044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.832133] env[65107]: INFO nova.compute.manager [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Terminating instance [ 736.849498] env[65107]: DEBUG oslo_vmware.api [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Task: {'id': task-5102576, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244531} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.853130] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 736.853130] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 736.853310] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 736.853492] env[65107]: INFO nova.compute.manager [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Took 1.17 seconds to destroy the instance on the hypervisor. [ 736.853738] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 736.854216] env[65107]: DEBUG nova.compute.manager [-] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 736.854336] env[65107]: DEBUG nova.network.neutron [-] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 736.854572] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.860200] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.860517] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.916354] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.938095] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b479e3-c288-48fb-a028-a543ab7b5d1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.948862] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8892013e-fffd-4123-9d98-a03482714863 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.994835] env[65107]: INFO nova.compute.manager [-] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Took 1.41 seconds to deallocate network for instance. [ 736.998314] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b021acfc-8a6f-4edf-9476-72d97f75165c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.014789] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53db918b-3b9c-4bc9-9599-1b8bd12ec05a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.040881] env[65107]: DEBUG nova.compute.provider_tree [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.050048] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102578, 'name': Rename_Task, 'duration_secs': 0.253321} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.050048] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 737.050048] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5803be6-b9b9-4c5d-97c3-445e4f14f2bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.059465] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 737.059465] env[65107]: value = "task-5102580" [ 737.059465] env[65107]: _type = "Task" [ 737.059465] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.075747] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102580, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.152273] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102579, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.157680] env[65107]: DEBUG nova.network.neutron [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Successfully created port: fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 737.172764] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.338297] env[65107]: DEBUG nova.compute.manager [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 737.338297] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.338297] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae5b2aa-f7a5-443f-97d8-24bc9dc63bd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.349767] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 737.350072] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-404370e0-499e-459e-befa-f6af624c099b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.358522] env[65107]: DEBUG oslo_vmware.api [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 737.358522] env[65107]: value = "task-5102581" [ 737.358522] env[65107]: _type = "Task" [ 737.358522] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.369916] env[65107]: DEBUG oslo_vmware.api [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.510047] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.545134] env[65107]: DEBUG nova.scheduler.client.report [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 737.550521] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 737.574733] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102580, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.588192] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:22:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a3c77cac-4e5e-476c-bee7-e5d40e2e12d5',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-991042522',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 737.589784] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 737.590194] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 737.590513] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 737.590843] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 737.591652] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 737.592280] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.592624] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 737.592865] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 737.593152] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 737.593426] env[65107]: DEBUG nova.virt.hardware [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 737.594825] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700ad7d8-5c48-4c96-bf9c-1bf1a4ba8ce6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.605813] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0b1c6c-f966-4787-b92e-92c83131d41b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.612237] env[65107]: DEBUG nova.compute.manager [req-6e2b8c3c-1e39-4f67-ba44-76f374344dd5 req-57abba07-7507-4621-a445-ae271cf3a12b service nova] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Received event network-vif-deleted-8226470b-4e4b-4b3f-b40f-b29d3149af69 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 737.651131] env[65107]: DEBUG oslo_vmware.api [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102579, 'name': PowerOnVM_Task, 'duration_secs': 0.694875} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.651432] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 737.651635] env[65107]: INFO nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Took 10.94 seconds to spawn the instance on the hypervisor. [ 737.651814] env[65107]: DEBUG nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 737.652638] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d91feb-09bd-4be8-bbbf-66f017aaecaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.772687] env[65107]: DEBUG nova.network.neutron [-] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 737.870029] env[65107]: DEBUG oslo_vmware.api [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102581, 'name': PowerOffVM_Task, 'duration_secs': 0.340472} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.870282] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 737.870449] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 737.870738] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-435ca0eb-24e4-421f-ac45-6d6445095b9f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.954737] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 737.954737] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 737.954737] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Deleting the datastore file [datastore2] 854f7a3f-9b7a-4506-8310-94533f0e4cfc {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 737.954737] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04a77be5-0f50-4725-9236-5e4dfcc7665a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.968019] env[65107]: DEBUG oslo_vmware.api [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for the task: (returnval){ [ 737.968019] env[65107]: value = "task-5102583" [ 737.968019] env[65107]: _type = "Task" [ 737.968019] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.980047] env[65107]: DEBUG oslo_vmware.api [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102583, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.054118] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.054866] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 738.059117] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.190s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.059117] env[65107]: DEBUG nova.objects.instance [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lazy-loading 'resources' on Instance uuid 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.074024] env[65107]: DEBUG oslo_vmware.api [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102580, 'name': PowerOnVM_Task, 'duration_secs': 0.957866} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.074307] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 738.074501] env[65107]: INFO nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Took 6.03 seconds to spawn the instance on the hypervisor. [ 738.074669] env[65107]: DEBUG nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 738.075590] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e41d17c-0015-4d23-a9f9-cd07d8007b04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.177898] env[65107]: INFO nova.compute.manager [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Took 44.26 seconds to build instance. [ 738.275471] env[65107]: INFO nova.compute.manager [-] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Took 1.42 seconds to deallocate network for instance. [ 738.439094] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "ad53146e-210d-4321-b819-90795c1af260" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.439436] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "ad53146e-210d-4321-b819-90795c1af260" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.439656] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "ad53146e-210d-4321-b819-90795c1af260-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.439838] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "ad53146e-210d-4321-b819-90795c1af260-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.440077] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "ad53146e-210d-4321-b819-90795c1af260-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.442627] env[65107]: INFO nova.compute.manager [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Terminating instance [ 738.479090] env[65107]: DEBUG oslo_vmware.api [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Task: {'id': task-5102583, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165616} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.479255] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.479428] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 738.479619] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.479812] env[65107]: INFO nova.compute.manager [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 738.480107] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 738.480325] env[65107]: DEBUG nova.compute.manager [-] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 738.480427] env[65107]: DEBUG nova.network.neutron [-] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 738.480679] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.481243] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.481506] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.552374] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.562282] env[65107]: DEBUG nova.compute.utils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 738.563904] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 738.567623] env[65107]: DEBUG nova.network.neutron [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 738.567623] env[65107]: WARNING neutronclient.v2_0.client [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.567623] env[65107]: WARNING neutronclient.v2_0.client [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.567623] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.567623] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.599359] env[65107]: INFO nova.compute.manager [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Took 39.72 seconds to build instance. [ 738.680750] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db294772-424d-484c-a756-830af9cb97ac tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "d40aec5a-d898-42f3-a121-503cf043f72b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.468s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.782937] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.798305] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "interface-d40aec5a-d898-42f3-a121-503cf043f72b-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.798529] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "interface-d40aec5a-d898-42f3-a121-503cf043f72b-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.798819] env[65107]: DEBUG nova.objects.instance [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lazy-loading 'flavor' on Instance uuid d40aec5a-d898-42f3-a121-503cf043f72b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.805626] env[65107]: DEBUG nova.network.neutron [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Successfully updated port: fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 738.840131] env[65107]: DEBUG nova.policy [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '241bae3dd1bc4f6198ffb030a938019e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71cd673a320348a9925e8f59637afc17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 738.902546] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 738.902931] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992689', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'name': 'volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c7739dfd-88d6-46ac-b3a0-e7dfe9b00332', 'attached_at': '', 'detached_at': '', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'serial': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 738.903889] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8837e153-f91b-4731-afcb-686acf3a2832 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.927397] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c577a9c-ee4b-41ef-b839-59a74214754a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.949617] env[65107]: DEBUG nova.compute.manager [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 738.949846] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.957424] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc/volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.960872] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b03ddaf-bc91-4c24-a217-f847e261a30d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.963586] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-976817fd-058f-4800-81a1-414094a76d10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.984690] env[65107]: DEBUG oslo_vmware.api [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 738.984690] env[65107]: value = "task-5102584" [ 738.984690] env[65107]: _type = "Task" [ 738.984690] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.987071] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 738.992209] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd749e50-627e-4c07-a5e3-75a7869ff23b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.005450] env[65107]: DEBUG oslo_vmware.api [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 739.005450] env[65107]: value = "task-5102585" [ 739.005450] env[65107]: _type = "Task" [ 739.005450] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.005962] env[65107]: DEBUG oslo_vmware.api [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102584, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.014871] env[65107]: DEBUG oslo_vmware.api [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.081965] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 739.101171] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a22041e-ce23-458e-b6f8-16b56ae0b347 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.797s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.178021] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b569ac1e-1b82-467d-83ef-f2080d72d82b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.187496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d89544-7f7a-4623-84dd-7b914bd69ba5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.191369] env[65107]: DEBUG nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 739.230948] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2d1f23-c897-458c-9334-eef4f188c95d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.240523] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844e91b3-4dfe-4183-97ec-e3f4c0ddc24e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.249774] env[65107]: DEBUG nova.network.neutron [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Successfully created port: 33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 739.261305] env[65107]: DEBUG nova.compute.provider_tree [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.300412] env[65107]: DEBUG nova.network.neutron [-] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 739.302919] env[65107]: WARNING neutronclient.v2_0.client [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.304215] env[65107]: WARNING openstack [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.304755] env[65107]: WARNING openstack [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.311267] env[65107]: DEBUG nova.objects.instance [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lazy-loading 'pci_requests' on Instance uuid d40aec5a-d898-42f3-a121-503cf043f72b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.316432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.316432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.316432] env[65107]: DEBUG nova.network.neutron [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 739.498210] env[65107]: DEBUG oslo_vmware.api [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102584, 'name': ReconfigVM_Task, 'duration_secs': 0.434175} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.498591] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Reconfigured VM instance instance-00000020 to attach disk [datastore2] volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc/volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.503502] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-038a728d-ebb2-4cee-a707-b0cffab89c6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.523506] env[65107]: DEBUG oslo_vmware.api [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102585, 'name': PowerOffVM_Task, 'duration_secs': 0.194333} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.524907] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 739.525125] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 739.525610] env[65107]: DEBUG oslo_vmware.api [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 739.525610] env[65107]: value = "task-5102586" [ 739.525610] env[65107]: _type = "Task" [ 739.525610] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.525735] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcf9bb09-3a39-4ff9-a480-cb8ec3ab97c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.538041] env[65107]: DEBUG oslo_vmware.api [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102586, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.589069] env[65107]: INFO nova.virt.block_device [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Booting with volume f04439cd-6130-4943-97ee-01d8501c2efc at /dev/sda [ 739.605409] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 739.605836] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 739.606169] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleting the datastore file [datastore1] ad53146e-210d-4321-b819-90795c1af260 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.606601] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e36ed6c9-96ed-4cc3-b447-d579ababc87d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.617022] env[65107]: DEBUG oslo_vmware.api [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 739.617022] env[65107]: value = "task-5102588" [ 739.617022] env[65107]: _type = "Task" [ 739.617022] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.628945] env[65107]: DEBUG oslo_vmware.api [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102588, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.630630] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9a02b19-a5d4-4f80-9a9c-f9dafefb0e4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.640020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a251e9be-cd33-405a-a4e2-e81a8de06a2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.680458] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-72cdc090-2a14-4b1c-b08b-5d4dee5bdfea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.690250] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791b256d-0ad4-4a76-9a58-439e57599346 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.731616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.732449] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d9466d-cc33-45bd-99b1-d1953770d311 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.739729] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb4f33f-ca29-40fc-881e-febc2142b7ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.755960] env[65107]: DEBUG nova.virt.block_device [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updating existing volume attachment record: 4cf52869-adcb-49ae-92be-8456bf35f64b {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 739.767214] env[65107]: DEBUG nova.scheduler.client.report [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.795573] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "be518c1d-edd7-40ff-b7cc-3310885b07b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.795838] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.796126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "be518c1d-edd7-40ff-b7cc-3310885b07b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.796272] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.796453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.800276] env[65107]: INFO nova.compute.manager [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Terminating instance [ 739.813865] env[65107]: INFO nova.compute.manager [-] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Took 1.33 seconds to deallocate network for instance. [ 739.817759] env[65107]: DEBUG nova.objects.base [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 739.817956] env[65107]: DEBUG nova.network.neutron [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 739.818262] env[65107]: WARNING neutronclient.v2_0.client [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.818556] env[65107]: WARNING neutronclient.v2_0.client [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.819135] env[65107]: WARNING openstack [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.819466] env[65107]: WARNING openstack [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.827670] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.828061] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.940312] env[65107]: DEBUG nova.network.neutron [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 740.024335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af207565-780b-4239-bd25-611609fbf9c5 tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "interface-d40aec5a-d898-42f3-a121-503cf043f72b-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.226s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.037710] env[65107]: DEBUG oslo_vmware.api [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102586, 'name': ReconfigVM_Task, 'duration_secs': 0.178836} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.039653] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.040012] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.046564] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992689', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'name': 'volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c7739dfd-88d6-46ac-b3a0-e7dfe9b00332', 'attached_at': '', 'detached_at': '', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'serial': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 740.079268] env[65107]: DEBUG nova.compute.manager [req-7bd272c3-b81f-471a-9c11-2372012d60e2 req-aff7bcbf-f76a-4c83-9f44-ca9275463204 service nova] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Received event network-vif-deleted-41ad83e5-4645-4e91-a223-473eefb69c38 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 740.116826] env[65107]: WARNING neutronclient.v2_0.client [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.117621] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.118016] env[65107]: WARNING openstack [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.136019] env[65107]: DEBUG oslo_vmware.api [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102588, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188782} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.136326] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.136513] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 740.136706] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.136888] env[65107]: INFO nova.compute.manager [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: ad53146e-210d-4321-b819-90795c1af260] Took 1.19 seconds to destroy the instance on the hypervisor. [ 740.137197] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 740.137426] env[65107]: DEBUG nova.compute.manager [-] [instance: ad53146e-210d-4321-b819-90795c1af260] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 740.137525] env[65107]: DEBUG nova.network.neutron [-] [instance: ad53146e-210d-4321-b819-90795c1af260] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 740.137786] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.138362] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.138655] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.232250] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.267137] env[65107]: DEBUG nova.network.neutron [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.271033] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.213s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.273882] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.704s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.273882] env[65107]: DEBUG nova.objects.instance [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lazy-loading 'resources' on Instance uuid 07b83b81-9c65-4180-b208-c96525b78d8a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.304646] env[65107]: DEBUG nova.compute.manager [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 740.304935] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.306385] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1a2647-e4bc-4fda-8837-29a13f08dbce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.310796] env[65107]: INFO nova.scheduler.client.report [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted allocations for instance 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6 [ 740.319287] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 740.320288] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6f62a01-2dc4-4e80-a4a4-5da3c37cb43b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.329782] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.334785] env[65107]: DEBUG oslo_vmware.api [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 740.334785] env[65107]: value = "task-5102589" [ 740.334785] env[65107]: _type = "Task" [ 740.334785] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.344878] env[65107]: DEBUG oslo_vmware.api [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.769731] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.770148] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Instance network_info: |[{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 740.770630] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:42:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcca9afe-edba-479c-93f6-1310627d946f', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.778404] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 740.781666] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.782184] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d6b732e-3f3d-4e37-a4ba-bb3bde7b63bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.815646] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.815646] env[65107]: value = "task-5102590" [ 740.815646] env[65107]: _type = "Task" [ 740.815646] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.829424] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102590, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.832917] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48200f5a-037a-45d4-9b6e-78ddf7924c6d tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.085s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.843864] env[65107]: DEBUG oslo_vmware.api [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102589, 'name': PowerOffVM_Task, 'duration_secs': 0.349817} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.844153] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 740.844386] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 740.847365] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b29f69c3-783c-4d02-9f4e-2655c095ab47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.865678] env[65107]: DEBUG nova.network.neutron [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Successfully updated port: 33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 740.925821] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 740.927295] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 740.927295] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Deleting the datastore file [datastore1] be518c1d-edd7-40ff-b7cc-3310885b07b4 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.930026] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c378007-f1f0-4ff5-b550-089978f282e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.940809] env[65107]: DEBUG oslo_vmware.api [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 740.940809] env[65107]: value = "task-5102592" [ 740.940809] env[65107]: _type = "Task" [ 740.940809] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.953482] env[65107]: DEBUG oslo_vmware.api [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102592, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.010428] env[65107]: DEBUG nova.network.neutron [-] [instance: ad53146e-210d-4321-b819-90795c1af260] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 741.107143] env[65107]: DEBUG nova.objects.instance [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'flavor' on Instance uuid c7739dfd-88d6-46ac-b3a0-e7dfe9b00332 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 741.329332] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102590, 'name': CreateVM_Task, 'duration_secs': 0.486872} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.329837] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 741.330643] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2411e2-1edb-4953-8bdc-b2c6b9e32654 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.333687] env[65107]: WARNING neutronclient.v2_0.client [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.334094] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.334292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.334928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 741.334928] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c1abe5e-8cc7-4946-a41a-d32fdb16638d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.342542] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 741.342542] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7660e-c1d3-4db6-d467-becea107d135" [ 741.342542] env[65107]: _type = "Task" [ 741.342542] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.343871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e842c21-2b27-4289-820b-89929efbe48c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.357415] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7660e-c1d3-4db6-d467-becea107d135, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.386409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.386659] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquired lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.386795] env[65107]: DEBUG nova.network.neutron [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 741.392202] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6905b5ff-9d07-4c07-b2a9-97ce4646bf14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.403106] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b17e37-abf2-4c86-8320-6792f1d9381b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.423532] env[65107]: DEBUG nova.compute.provider_tree [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 741.452186] env[65107]: DEBUG oslo_vmware.api [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102592, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161028} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.452527] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 741.452681] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 741.452896] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.453121] env[65107]: INFO nova.compute.manager [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 741.453388] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 741.453672] env[65107]: DEBUG nova.compute.manager [-] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 741.453755] env[65107]: DEBUG nova.network.neutron [-] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 741.453958] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.454544] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.454802] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.514054] env[65107]: INFO nova.compute.manager [-] [instance: ad53146e-210d-4321-b819-90795c1af260] Took 1.38 seconds to deallocate network for instance. [ 741.529695] env[65107]: DEBUG nova.compute.manager [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Received event network-vif-plugged-fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 741.530011] env[65107]: DEBUG oslo_concurrency.lockutils [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Acquiring lock "369caf24-48a6-4640-8601-e30fd47453da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.530194] env[65107]: DEBUG oslo_concurrency.lockutils [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Lock "369caf24-48a6-4640-8601-e30fd47453da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.530385] env[65107]: DEBUG oslo_concurrency.lockutils [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Lock "369caf24-48a6-4640-8601-e30fd47453da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.530846] env[65107]: DEBUG nova.compute.manager [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] No waiting events found dispatching network-vif-plugged-fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 741.530846] env[65107]: WARNING nova.compute.manager [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Received unexpected event network-vif-plugged-fcca9afe-edba-479c-93f6-1310627d946f for instance with vm_state building and task_state spawning. [ 741.531050] env[65107]: DEBUG nova.compute.manager [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Received event network-changed-fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 741.531227] env[65107]: DEBUG nova.compute.manager [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Refreshing instance network info cache due to event network-changed-fcca9afe-edba-479c-93f6-1310627d946f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 741.531389] env[65107]: DEBUG oslo_concurrency.lockutils [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Acquiring lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.531520] env[65107]: DEBUG oslo_concurrency.lockutils [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Acquired lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.531674] env[65107]: DEBUG nova.network.neutron [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Refreshing network info cache for port fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 741.550558] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.615046] env[65107]: DEBUG oslo_concurrency.lockutils [None req-27129380-4218-4d57-aca1-b8afbff05ffe tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.868s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.854131] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7660e-c1d3-4db6-d467-becea107d135, 'name': SearchDatastore_Task, 'duration_secs': 0.026437} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.854496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.854698] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 741.854938] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.855094] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.855312] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 741.855718] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2587142-5714-43ec-a91d-0a6706144bc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.865653] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 741.865848] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 741.866708] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38182154-c7c5-46d0-a93d-d890ae332dbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.873219] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 741.873219] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52496192-2521-e005-33f6-d419bbe65de0" [ 741.873219] env[65107]: _type = "Task" [ 741.873219] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.882011] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52496192-2521-e005-33f6-d419bbe65de0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.892997] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.893548] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.902126] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 741.902657] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 741.902868] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 741.903045] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 741.903398] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 741.903398] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 741.903527] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 741.903797] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.903870] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 741.904043] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 741.904204] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 741.904371] env[65107]: DEBUG nova.virt.hardware [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 741.907068] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fe00c8-ea56-4e32-9547-b5fefd6628c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.916921] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e4051e-5000-4d93-904d-b5be70b749fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.959234] env[65107]: ERROR nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] [req-e1bb9390-da9e-400a-abce-804a17b6da3c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e1bb9390-da9e-400a-abce-804a17b6da3c"}]} [ 741.971025] env[65107]: DEBUG nova.network.neutron [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 741.983678] env[65107]: DEBUG nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 741.999426] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.002204] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.008654] env[65107]: DEBUG nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 742.008871] env[65107]: DEBUG nova.compute.provider_tree [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 742.022543] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.025159] env[65107]: DEBUG nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 742.034322] env[65107]: WARNING neutronclient.v2_0.client [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.034919] env[65107]: WARNING openstack [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.035708] env[65107]: WARNING openstack [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.050541] env[65107]: DEBUG nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 742.095551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.095792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.190465] env[65107]: WARNING neutronclient.v2_0.client [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.190710] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.192048] env[65107]: WARNING openstack [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.230299] env[65107]: WARNING openstack [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.230567] env[65107]: WARNING openstack [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.270820] env[65107]: DEBUG nova.compute.manager [None req-5c331ae6-84aa-4ecd-b031-419c03a736d1 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 742.270820] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9c1ee2-ecca-4040-8a47-955fd576b296 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.326505] env[65107]: DEBUG nova.network.neutron [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updating instance_info_cache with network_info: [{"id": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "address": "fa:16:3e:27:e4:ae", "network": {"id": "472c4d13-7d57-4d84-b35d-7f05f937120a", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1598181546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71cd673a320348a9925e8f59637afc17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33799b3d-5a", "ovs_interfaceid": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.357675] env[65107]: WARNING neutronclient.v2_0.client [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.359341] env[65107]: WARNING openstack [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.359341] env[65107]: WARNING openstack [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.389737] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52496192-2521-e005-33f6-d419bbe65de0, 'name': SearchDatastore_Task, 'duration_secs': 0.011926} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.391275] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd7518b7-1902-452f-be6f-f692b3a53d9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.402486] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 742.402486] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5258306f-15a3-10d5-26ef-4a60a97d565e" [ 742.402486] env[65107]: _type = "Task" [ 742.402486] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.416230] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5258306f-15a3-10d5-26ef-4a60a97d565e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.464999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.464999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.464999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.464999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.465663] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.468754] env[65107]: INFO nova.compute.manager [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Terminating instance [ 742.477328] env[65107]: DEBUG nova.network.neutron [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updated VIF entry in instance network info cache for port fcca9afe-edba-479c-93f6-1310627d946f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 742.477782] env[65107]: DEBUG nova.network.neutron [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.598027] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 742.623423] env[65107]: DEBUG nova.network.neutron [-] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.714138] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2499867f-a15c-47e5-8382-ce2f08691a10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.722282] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b82684d-1335-43d8-a470-278f5ddceb42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.762098] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9594cdc1-9d88-4f94-88d2-0ed527ed2dfb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.771676] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ce6839-48cd-401a-a377-b8e6c714eee6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.786730] env[65107]: DEBUG nova.compute.provider_tree [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 742.788691] env[65107]: INFO nova.compute.manager [None req-5c331ae6-84aa-4ecd-b031-419c03a736d1 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] instance snapshotting [ 742.789385] env[65107]: DEBUG nova.objects.instance [None req-5c331ae6-84aa-4ecd-b031-419c03a736d1 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lazy-loading 'flavor' on Instance uuid c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 742.834572] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Releasing lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.835205] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance network_info: |[{"id": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "address": "fa:16:3e:27:e4:ae", "network": {"id": "472c4d13-7d57-4d84-b35d-7f05f937120a", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1598181546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71cd673a320348a9925e8f59637afc17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33799b3d-5a", "ovs_interfaceid": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 742.835703] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:e4:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33799b3d-5ade-4cb0-ae83-3d00bfd972ca', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.843511] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Creating folder: Project (71cd673a320348a9925e8f59637afc17). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.844168] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6fb85ee-9d39-4a02-971c-06bc4397a090 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.857039] env[65107]: DEBUG nova.compute.manager [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Received event network-vif-deleted-454bd66e-49a9-4417-869d-fb1c0f08d472 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 742.857363] env[65107]: DEBUG nova.compute.manager [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Received event network-vif-plugged-33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 742.857473] env[65107]: DEBUG oslo_concurrency.lockutils [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Acquiring lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.858037] env[65107]: DEBUG oslo_concurrency.lockutils [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.858037] env[65107]: DEBUG oslo_concurrency.lockutils [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.858037] env[65107]: DEBUG nova.compute.manager [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] No waiting events found dispatching network-vif-plugged-33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 742.858189] env[65107]: WARNING nova.compute.manager [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Received unexpected event network-vif-plugged-33799b3d-5ade-4cb0-ae83-3d00bfd972ca for instance with vm_state building and task_state spawning. [ 742.858853] env[65107]: DEBUG nova.compute.manager [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Received event network-changed-33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 742.858853] env[65107]: DEBUG nova.compute.manager [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Refreshing instance network info cache due to event network-changed-33799b3d-5ade-4cb0-ae83-3d00bfd972ca. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 742.858853] env[65107]: DEBUG oslo_concurrency.lockutils [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Acquiring lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.858853] env[65107]: DEBUG oslo_concurrency.lockutils [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Acquired lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.859286] env[65107]: DEBUG nova.network.neutron [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Refreshing network info cache for port 33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 742.861322] env[65107]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 742.861910] env[65107]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65107) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 742.862193] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Folder already exists: Project (71cd673a320348a9925e8f59637afc17). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 742.862308] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Creating folder: Instances. Parent ref: group-v992644. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.862538] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3aab1399-3810-49f2-bbc2-9c59f870d09f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.873854] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Created folder: Instances in parent group-v992644. [ 742.874704] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 742.874704] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 742.874873] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a47e6d7-9997-4450-aff8-73c476b63bb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.895216] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.895216] env[65107]: value = "task-5102595" [ 742.895216] env[65107]: _type = "Task" [ 742.895216] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.904157] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102595, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.915333] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5258306f-15a3-10d5-26ef-4a60a97d565e, 'name': SearchDatastore_Task, 'duration_secs': 0.012399} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.915333] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.915333] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 742.915554] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3862b30b-da6f-4fd5-8f52-9a48e710e6b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.918175] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "d40aec5a-d898-42f3-a121-503cf043f72b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.918426] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "d40aec5a-d898-42f3-a121-503cf043f72b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.918664] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "d40aec5a-d898-42f3-a121-503cf043f72b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.918874] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "d40aec5a-d898-42f3-a121-503cf043f72b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.919521] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "d40aec5a-d898-42f3-a121-503cf043f72b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.921792] env[65107]: INFO nova.compute.manager [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Terminating instance [ 742.929340] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 742.929340] env[65107]: value = "task-5102596" [ 742.929340] env[65107]: _type = "Task" [ 742.929340] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.943952] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.973683] env[65107]: DEBUG nova.compute.manager [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 742.973846] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 742.974738] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a228ea9-fca4-454b-811f-638353b8bd25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.985667] env[65107]: DEBUG oslo_concurrency.lockutils [req-c8ec561e-bcc3-4ce8-b20c-92e014a45460 req-4de9f08e-0cf0-4bd3-b0b2-88a4c40cbcae service nova] Releasing lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.986484] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 742.986484] env[65107]: value = "task-5102597" [ 742.986484] env[65107]: _type = "Task" [ 742.986484] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.997721] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102597, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.132437] env[65107]: INFO nova.compute.manager [-] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Took 1.68 seconds to deallocate network for instance. [ 743.133992] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.136371] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.136852] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.137162] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.137433] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.137632] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.141593] env[65107]: INFO nova.compute.manager [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Terminating instance [ 743.299204] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bba7aa0-0998-4ffa-a87b-329b961a2052 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.320325] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bd25d0-ab5d-4185-8eba-8a018e7955ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.334089] env[65107]: DEBUG nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 63 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 743.335465] env[65107]: DEBUG nova.compute.provider_tree [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 63 to 64 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 743.335872] env[65107]: DEBUG nova.compute.provider_tree [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 743.368432] env[65107]: WARNING neutronclient.v2_0.client [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.369678] env[65107]: WARNING openstack [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.370215] env[65107]: WARNING openstack [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.414675] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102595, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.431512] env[65107]: DEBUG nova.compute.manager [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 743.432756] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.432863] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455cc0c8-48a0-4489-83c2-c3bf14b1a741 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.455822] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 743.455822] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102596, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.455822] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8232f30-e5e0-46fc-b537-90c4d2fbfbf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.469952] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 743.469952] env[65107]: value = "task-5102598" [ 743.469952] env[65107]: _type = "Task" [ 743.469952] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.480942] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.485547] env[65107]: WARNING openstack [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.485724] env[65107]: WARNING openstack [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.506413] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102597, 'name': PowerOffVM_Task, 'duration_secs': 0.334435} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.506693] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.506895] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 743.507106] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992689', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'name': 'volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c7739dfd-88d6-46ac-b3a0-e7dfe9b00332', 'attached_at': '', 'detached_at': '', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'serial': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 743.508037] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde62597-47ec-4458-a8b3-e0a793a4b937 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.538620] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb324585-7377-44be-a60e-c613d511ae58 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.546249] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c889d42d-9d5d-4939-9880-33defd151035 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.570326] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc83401-7840-4919-b79f-0ce9130f7f26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.575463] env[65107]: WARNING neutronclient.v2_0.client [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.576443] env[65107]: WARNING openstack [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.576946] env[65107]: WARNING openstack [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.598896] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The volume has not been displaced from its original location: [datastore2] volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc/volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 743.604268] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Reconfiguring VM instance instance-00000020 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 743.604679] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0425ac2-417b-4c62-9c67-9c0ecad6c65e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.625369] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 743.625369] env[65107]: value = "task-5102599" [ 743.625369] env[65107]: _type = "Task" [ 743.625369] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.644041] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102599, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.654291] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.658422] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "refresh_cache-c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.658422] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquired lock "refresh_cache-c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.658422] env[65107]: DEBUG nova.network.neutron [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 743.683056] env[65107]: DEBUG nova.network.neutron [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updated VIF entry in instance network info cache for port 33799b3d-5ade-4cb0-ae83-3d00bfd972ca. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 743.684321] env[65107]: DEBUG nova.network.neutron [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updating instance_info_cache with network_info: [{"id": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "address": "fa:16:3e:27:e4:ae", "network": {"id": "472c4d13-7d57-4d84-b35d-7f05f937120a", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1598181546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71cd673a320348a9925e8f59637afc17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33799b3d-5a", "ovs_interfaceid": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.842307] env[65107]: DEBUG nova.compute.manager [None req-5c331ae6-84aa-4ecd-b031-419c03a736d1 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Instance disappeared during snapshot {{(pid=65107) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 743.845499] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.572s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.847958] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.902s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.848202] env[65107]: DEBUG nova.objects.instance [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 743.875338] env[65107]: INFO nova.scheduler.client.report [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Deleted allocations for instance 07b83b81-9c65-4180-b208-c96525b78d8a [ 743.908076] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102595, 'name': CreateVM_Task, 'duration_secs': 0.679764} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.908076] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 743.908076] env[65107]: WARNING neutronclient.v2_0.client [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.908500] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992651', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'name': 'volume-f04439cd-6130-4943-97ee-01d8501c2efc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd65e98bb-f54d-4803-95ca-0fc9b6c184bc', 'attached_at': '', 'detached_at': '', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'serial': 'f04439cd-6130-4943-97ee-01d8501c2efc'}, 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'attachment_id': '4cf52869-adcb-49ae-92be-8456bf35f64b', 'mount_device': '/dev/sda', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=65107) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 743.908640] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Root volume attach. Driver type: vmdk {{(pid=65107) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 743.909469] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77923abb-eecc-4115-b792-dc35595c2729 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.918929] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06ea078-4b42-4104-ad17-85fbd9a6c52f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.926670] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89edd14-4460-4015-ab5e-a024cdefee94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.938129] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-f06bfa64-2c49-4b17-8949-ba039ff6f8ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.947053] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568281} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.949520] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 743.950015] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.950086] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 743.950086] env[65107]: value = "task-5102600" [ 743.950086] env[65107]: _type = "Task" [ 743.950086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.951495] env[65107]: DEBUG nova.compute.manager [req-9e63e37c-d7b4-45e4-8274-5651c91aa000 req-23a9bbbf-e092-41db-97f2-8cf5f8a77f08 service nova] [instance: ad53146e-210d-4321-b819-90795c1af260] Received event network-vif-deleted-8db11d31-614c-4e71-bb14-c5327b541437 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 743.951719] env[65107]: DEBUG nova.compute.manager [req-9e63e37c-d7b4-45e4-8274-5651c91aa000 req-23a9bbbf-e092-41db-97f2-8cf5f8a77f08 service nova] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Received event network-vif-deleted-3bb17344-0031-4226-9117-1366ecf93330 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 743.952045] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-678671da-a620-4794-b36c-54d6ef256de9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.964255] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102600, 'name': RelocateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.966243] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 743.966243] env[65107]: value = "task-5102601" [ 743.966243] env[65107]: _type = "Task" [ 743.966243] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.978852] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102601, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.984520] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102598, 'name': PowerOffVM_Task, 'duration_secs': 0.323673} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.987876] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.987943] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 743.988569] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ee5b573-92fe-482d-a194-90249e1b2793 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.029935] env[65107]: DEBUG nova.compute.manager [None req-5c331ae6-84aa-4ecd-b031-419c03a736d1 tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Found 0 images (rotation: 2) {{(pid=65107) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 744.070873] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 744.071276] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 744.071605] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Deleting the datastore file [datastore2] d40aec5a-d898-42f3-a121-503cf043f72b {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 744.071988] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2242853-bd42-4891-904d-c2799739eabc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.080948] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for the task: (returnval){ [ 744.080948] env[65107]: value = "task-5102603" [ 744.080948] env[65107]: _type = "Task" [ 744.080948] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.092219] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.136478] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102599, 'name': ReconfigVM_Task, 'duration_secs': 0.244062} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.137507] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Reconfigured VM instance instance-00000020 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 744.142080] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a44d247-4bbb-4edd-b316-1a46f37c56f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.159627] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 744.159627] env[65107]: value = "task-5102604" [ 744.159627] env[65107]: _type = "Task" [ 744.159627] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.164911] env[65107]: WARNING neutronclient.v2_0.client [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.165681] env[65107]: WARNING openstack [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.166032] env[65107]: WARNING openstack [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.185993] env[65107]: DEBUG oslo_concurrency.lockutils [req-ff90ca7a-243c-4bef-9444-1a121cef3aee req-8ddc6527-b599-4bb7-a715-3e09095467e4 service nova] Releasing lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.186416] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102604, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.202084] env[65107]: DEBUG nova.network.neutron [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 744.266083] env[65107]: DEBUG nova.network.neutron [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 744.382799] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b19ffea-90d2-443f-9b9c-34a51a1a37da tempest-MultipleCreateTestJSON-827681759 tempest-MultipleCreateTestJSON-827681759-project-member] Lock "07b83b81-9c65-4180-b208-c96525b78d8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.162s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.465015] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102600, 'name': RelocateVM_Task} progress is 19%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.475672] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131676} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.475964] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.477057] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42ad0c3-ed6b-44fa-a495-dfbd11c1f94c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.502475] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.502656] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-055c2797-caf3-4198-a0ba-045f7ee14dc4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.525514] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 744.525514] env[65107]: value = "task-5102605" [ 744.525514] env[65107]: _type = "Task" [ 744.525514] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.535978] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.595405] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.671497] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102604, 'name': ReconfigVM_Task, 'duration_secs': 0.444858} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.671836] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992689', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'name': 'volume-9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c7739dfd-88d6-46ac-b3a0-e7dfe9b00332', 'attached_at': '', 'detached_at': '', 'volume_id': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc', 'serial': '9fb5ce21-7b37-45d3-bfd4-2f96beaf81bc'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 744.672178] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.673070] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163fe0ee-2af5-4279-b523-43b413f95c44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.683508] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 744.683508] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-172645de-7471-4a35-9086-08968474fc30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.768390] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Releasing lock "refresh_cache-c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.768890] env[65107]: DEBUG nova.compute.manager [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 744.769156] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.770075] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d89fdb0-613b-40ee-a9d2-8070b875d568 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.778982] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 744.779166] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c70ece5b-5a5d-4b6d-9cf4-b288e9408696 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.786874] env[65107]: DEBUG oslo_vmware.api [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 744.786874] env[65107]: value = "task-5102607" [ 744.786874] env[65107]: _type = "Task" [ 744.786874] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.797085] env[65107]: DEBUG oslo_vmware.api [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.864434] env[65107]: DEBUG oslo_concurrency.lockutils [None req-227a0a62-b0ba-43a0-9ccb-592975e9382e tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.865997] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.040s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.868508] env[65107]: INFO nova.compute.claims [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.967502] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102600, 'name': RelocateVM_Task} progress is 20%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.039793] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102605, 'name': ReconfigVM_Task, 'duration_secs': 0.338203} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.040294] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.041317] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2dcef674-37e5-4434-9e98-28dbc44f608a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.051153] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 745.051153] env[65107]: value = "task-5102608" [ 745.051153] env[65107]: _type = "Task" [ 745.051153] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.063908] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102608, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.091823] env[65107]: DEBUG oslo_vmware.api [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Task: {'id': task-5102603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.538392} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.092164] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.092448] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.093103] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.093103] env[65107]: INFO nova.compute.manager [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Took 1.66 seconds to destroy the instance on the hypervisor. [ 745.093103] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.093390] env[65107]: DEBUG nova.compute.manager [-] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 745.093440] env[65107]: DEBUG nova.network.neutron [-] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 745.093761] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.094254] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.094480] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.144286] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.297682] env[65107]: DEBUG oslo_vmware.api [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102607, 'name': PowerOffVM_Task, 'duration_secs': 0.244156} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.297682] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 745.298124] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 745.298124] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-216b6455-c317-4ad8-a6f1-ea7502da328b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.324768] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 745.325018] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 745.325300] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Deleting the datastore file [datastore2] c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 745.325696] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4d02a39-0b8c-49d7-96c3-62d46bdb1928 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.333146] env[65107]: DEBUG oslo_vmware.api [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for the task: (returnval){ [ 745.333146] env[65107]: value = "task-5102610" [ 745.333146] env[65107]: _type = "Task" [ 745.333146] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.342087] env[65107]: DEBUG oslo_vmware.api [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102610, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.453372] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 745.453372] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 745.453372] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleting the datastore file [datastore1] c7739dfd-88d6-46ac-b3a0-e7dfe9b00332 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 745.453675] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-695025bc-b217-4d37-ba5c-d0adc4a63b5c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.464776] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102600, 'name': RelocateVM_Task, 'duration_secs': 1.109968} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.467899] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 745.467899] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992651', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'name': 'volume-f04439cd-6130-4943-97ee-01d8501c2efc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd65e98bb-f54d-4803-95ca-0fc9b6c184bc', 'attached_at': '', 'detached_at': '', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'serial': 'f04439cd-6130-4943-97ee-01d8501c2efc'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 745.467899] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 745.467899] env[65107]: value = "task-5102611" [ 745.467899] env[65107]: _type = "Task" [ 745.467899] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.467899] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6182579-5eb6-4b6c-935d-cc6b4ea5d9aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.483455] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.498267] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bda61d-b3b0-4a69-93d3-25c3bf634947 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.530097] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] volume-f04439cd-6130-4943-97ee-01d8501c2efc/volume-f04439cd-6130-4943-97ee-01d8501c2efc.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.530097] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8bd2a3a-df87-4bd6-9472-746cdd879a44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.553128] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 745.553128] env[65107]: value = "task-5102612" [ 745.553128] env[65107]: _type = "Task" [ 745.553128] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.567246] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102608, 'name': Rename_Task, 'duration_secs': 0.205505} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.571154] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 745.571322] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102612, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.571571] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de615f39-0f5d-4fe6-a181-da49f93400ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.579847] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 745.579847] env[65107]: value = "task-5102613" [ 745.579847] env[65107]: _type = "Task" [ 745.579847] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.589890] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.848984] env[65107]: DEBUG oslo_vmware.api [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Task: {'id': task-5102610, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.107396} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.849296] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.849474] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.849670] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.851292] env[65107]: INFO nova.compute.manager [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Took 1.08 seconds to destroy the instance on the hypervisor. [ 745.851292] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.851292] env[65107]: DEBUG nova.compute.manager [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 745.851292] env[65107]: DEBUG nova.network.neutron [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 745.851292] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.851292] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.851834] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.859302] env[65107]: DEBUG nova.network.neutron [-] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 745.886667] env[65107]: DEBUG nova.network.neutron [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 745.886911] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.984909] env[65107]: DEBUG oslo_vmware.api [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17357} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.985445] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.985735] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.985973] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.986198] env[65107]: INFO nova.compute.manager [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Took 3.01 seconds to destroy the instance on the hypervisor. [ 745.986544] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.986804] env[65107]: DEBUG nova.compute.manager [-] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 745.986914] env[65107]: DEBUG nova.network.neutron [-] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 745.987251] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.987976] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.988427] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.039000] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.073678] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102612, 'name': ReconfigVM_Task, 'duration_secs': 0.29817} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.073999] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Reconfigured VM instance instance-00000028 to attach disk [datastore1] volume-f04439cd-6130-4943-97ee-01d8501c2efc/volume-f04439cd-6130-4943-97ee-01d8501c2efc.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.082283] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90a83efd-40b1-439d-b18f-5c8c40a0c944 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.106557] env[65107]: DEBUG oslo_vmware.api [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102613, 'name': PowerOnVM_Task, 'duration_secs': 0.507994} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.112228] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 746.112665] env[65107]: INFO nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Took 8.56 seconds to spawn the instance on the hypervisor. [ 746.112747] env[65107]: DEBUG nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 746.114157] env[65107]: DEBUG nova.compute.manager [req-3ecdb5fb-070d-421f-bad9-4a08605cc887 req-b3beb1d9-5756-40e8-8b59-de3a7c9eefd7 service nova] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Received event network-vif-deleted-b3b4ec0d-49c0-4c0a-845c-157dd80604fb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 746.114971] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 746.114971] env[65107]: value = "task-5102614" [ 746.114971] env[65107]: _type = "Task" [ 746.114971] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.116093] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382c21af-26cd-4b92-a84c-6cecb19e97e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.136586] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102614, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.362630] env[65107]: INFO nova.compute.manager [-] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Took 1.27 seconds to deallocate network for instance. [ 746.391471] env[65107]: DEBUG nova.network.neutron [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.604732] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b66958-8cc8-4f86-9176-d50102e84939 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.612411] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0981d9-2221-49a8-91a6-7b000f0ec78a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.659202] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5dc6b7e-9230-4480-8611-028fb7f10563 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.665250] env[65107]: INFO nova.compute.manager [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Took 40.38 seconds to build instance. [ 746.673971] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102614, 'name': ReconfigVM_Task, 'duration_secs': 0.13894} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.674864] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d827130a-132c-4de5-9b46-eac3136c5cf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.680907] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992651', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'name': 'volume-f04439cd-6130-4943-97ee-01d8501c2efc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd65e98bb-f54d-4803-95ca-0fc9b6c184bc', 'attached_at': '', 'detached_at': '', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'serial': 'f04439cd-6130-4943-97ee-01d8501c2efc'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 746.682237] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9edaf6f-b573-47c9-a5ff-fe71b834c093 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.695672] env[65107]: DEBUG nova.compute.provider_tree [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.699541] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 746.699541] env[65107]: value = "task-5102615" [ 746.699541] env[65107]: _type = "Task" [ 746.699541] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.708688] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102615, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.848151] env[65107]: DEBUG nova.network.neutron [-] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.871147] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.897198] env[65107]: INFO nova.compute.manager [-] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Took 1.05 seconds to deallocate network for instance. [ 747.036060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.036355] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.036563] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.038042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.038042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.039763] env[65107]: INFO nova.compute.manager [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Terminating instance [ 747.168203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cddec31d-d80c-49f9-9924-c042a54c6f62 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.776s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.199187] env[65107]: DEBUG nova.scheduler.client.report [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 747.220755] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102615, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.352912] env[65107]: INFO nova.compute.manager [-] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Took 1.37 seconds to deallocate network for instance. [ 747.406766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.544599] env[65107]: DEBUG nova.compute.manager [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 747.544762] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.546276] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e31f60-7f34-4b76-8efa-f7424779ff72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.555329] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.556051] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-367ba473-f490-4275-91d7-21e430e17266 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.565300] env[65107]: DEBUG oslo_vmware.api [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 747.565300] env[65107]: value = "task-5102616" [ 747.565300] env[65107]: _type = "Task" [ 747.565300] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.579283] env[65107]: DEBUG oslo_vmware.api [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.715231] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.849s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.715883] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 747.718961] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102615, 'name': Rename_Task, 'duration_secs': 0.659287} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.722794] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.935s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.727025] env[65107]: INFO nova.compute.claims [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.730185] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 747.730774] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3e9d68e-4870-476f-8026-baad5b025a0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.743342] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 747.743342] env[65107]: value = "task-5102617" [ 747.743342] env[65107]: _type = "Task" [ 747.743342] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.754175] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.955324] env[65107]: INFO nova.compute.manager [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Took 0.60 seconds to detach 1 volumes for instance. [ 748.076206] env[65107]: DEBUG oslo_vmware.api [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102616, 'name': PowerOffVM_Task, 'duration_secs': 0.220423} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.076531] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 748.076689] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 748.076947] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2d1f950-5b66-4fa3-ae70-3825e7ae05b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.161216] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 748.161550] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 748.161618] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Deleting the datastore file [datastore2] b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.161891] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-470c9ef9-54da-4614-9efd-320ae1cdf91a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.170499] env[65107]: DEBUG oslo_vmware.api [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 748.170499] env[65107]: value = "task-5102619" [ 748.170499] env[65107]: _type = "Task" [ 748.170499] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.178861] env[65107]: DEBUG oslo_vmware.api [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102619, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.224874] env[65107]: DEBUG nova.compute.utils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 748.226782] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 748.227277] env[65107]: DEBUG nova.network.neutron [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 748.227841] env[65107]: WARNING neutronclient.v2_0.client [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.228286] env[65107]: WARNING neutronclient.v2_0.client [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.229104] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.229575] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.254391] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102617, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.317571] env[65107]: DEBUG nova.policy [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 748.464195] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.682095] env[65107]: DEBUG oslo_vmware.api [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102619, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.682742] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.682982] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.683185] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.683363] env[65107]: INFO nova.compute.manager [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 748.683644] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 748.683844] env[65107]: DEBUG nova.compute.manager [-] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 748.683899] env[65107]: DEBUG nova.network.neutron [-] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 748.684160] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.684683] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.684936] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.715636] env[65107]: DEBUG nova.network.neutron [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Successfully created port: b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 748.738279] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 748.764803] env[65107]: DEBUG oslo_vmware.api [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102617, 'name': PowerOnVM_Task, 'duration_secs': 0.67732} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.765070] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 748.765463] env[65107]: INFO nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Took 6.86 seconds to spawn the instance on the hypervisor. [ 748.765463] env[65107]: DEBUG nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 748.766370] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfffacd5-a9b3-4711-8761-9bc52a00ee4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.829705] env[65107]: DEBUG nova.compute.manager [req-ae4d02c5-3c6a-4b89-aa3f-d93b3c4b66d4 req-d79c8aaf-6337-4478-864c-d5caa3f231c6 service nova] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Received event network-vif-deleted-abf46aae-348e-4b4b-acfe-35830ba162e1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 748.864511] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.289927] env[65107]: INFO nova.compute.manager [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Took 41.56 seconds to build instance. [ 749.354916] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e899708-22aa-46b6-8fc2-8b7028cf2152 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.363259] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f51bf4c-aefd-45e7-a0da-01bdfd962bf2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.401612] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18b4512-1c44-402f-910c-26bf7f27b06a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.411494] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62f93f1-06f4-4bd7-83c9-180c00949140 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.426839] env[65107]: DEBUG nova.compute.provider_tree [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.750932] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 749.781738] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 749.781994] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 749.782166] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 749.782344] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 749.782487] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 749.782629] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 749.782917] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.783148] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 749.783328] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 749.783519] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 749.783655] env[65107]: DEBUG nova.virt.hardware [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 749.784576] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44eb9dd7-6496-440a-bb8a-f4c46e7c13f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.794196] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca3aa1c-612d-476c-a1cb-e7b0887b669c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.799541] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80cc2627-b323-4d58-8127-a9fb775d9c5f tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.104s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.929875] env[65107]: DEBUG nova.scheduler.client.report [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.940328] env[65107]: DEBUG nova.compute.manager [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 750.026422] env[65107]: DEBUG nova.network.neutron [-] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.437163] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.714s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.437606] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 750.441211] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.603s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.442668] env[65107]: INFO nova.compute.claims [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.466455] env[65107]: DEBUG nova.network.neutron [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Successfully updated port: b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 750.468986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.530663] env[65107]: INFO nova.compute.manager [-] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Took 1.85 seconds to deallocate network for instance. [ 750.952015] env[65107]: DEBUG nova.compute.utils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 750.953690] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 750.953826] env[65107]: DEBUG nova.network.neutron [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 750.954936] env[65107]: WARNING neutronclient.v2_0.client [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.954936] env[65107]: WARNING neutronclient.v2_0.client [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.955089] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.956126] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.974488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.974670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.975031] env[65107]: DEBUG nova.network.neutron [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 751.039247] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.088536] env[65107]: DEBUG nova.policy [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd468fcc45f624d83afc0a6ba7dc3f47d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1fd7913632a4a5b9e076b5b156fa5fa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 751.133496] env[65107]: DEBUG nova.compute.manager [req-1ee2f911-1367-4c13-9b28-7e0375886db4 req-ef51f719-c2cc-42b0-8530-d3fa261fa08d service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-vif-plugged-b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 751.133764] env[65107]: DEBUG oslo_concurrency.lockutils [req-1ee2f911-1367-4c13-9b28-7e0375886db4 req-ef51f719-c2cc-42b0-8530-d3fa261fa08d service nova] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.133937] env[65107]: DEBUG oslo_concurrency.lockutils [req-1ee2f911-1367-4c13-9b28-7e0375886db4 req-ef51f719-c2cc-42b0-8530-d3fa261fa08d service nova] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.134550] env[65107]: DEBUG oslo_concurrency.lockutils [req-1ee2f911-1367-4c13-9b28-7e0375886db4 req-ef51f719-c2cc-42b0-8530-d3fa261fa08d service nova] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.134550] env[65107]: DEBUG nova.compute.manager [req-1ee2f911-1367-4c13-9b28-7e0375886db4 req-ef51f719-c2cc-42b0-8530-d3fa261fa08d service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] No waiting events found dispatching network-vif-plugged-b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 751.134550] env[65107]: WARNING nova.compute.manager [req-1ee2f911-1367-4c13-9b28-7e0375886db4 req-ef51f719-c2cc-42b0-8530-d3fa261fa08d service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received unexpected event network-vif-plugged-b44410c9-c28d-4e13-879d-3e495fde9ff4 for instance with vm_state building and task_state spawning. [ 751.432662] env[65107]: DEBUG nova.compute.manager [req-a54d70f6-6fa9-4ac1-a54d-01a9d6e847dd req-55e70a48-a9ac-4139-a17a-1680992d447a service nova] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Received event network-vif-deleted-f9582d02-fbd0-490d-8f90-10bc9cf32af8 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 751.465489] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 751.481601] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.481601] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.689020] env[65107]: DEBUG nova.network.neutron [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Successfully created port: 0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 751.834778] env[65107]: DEBUG nova.network.neutron [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 751.936781] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.937182] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.031817] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9674e7b-8aa2-4e48-9298-21decbc6176c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.041372] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304d9f01-4362-49ca-a9ff-561b66fc8e20 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.074585] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5ae7e7-2598-4704-b4ec-d7fe7fb8bd0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.085063] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e978c2-16eb-4c4e-861e-a789ed163ab1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.098761] env[65107]: DEBUG nova.compute.provider_tree [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.479092] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 752.504914] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 752.505186] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 752.505347] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 752.505606] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 752.505821] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 752.505984] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 752.506208] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.506365] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 752.506609] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 752.506777] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 752.506948] env[65107]: DEBUG nova.virt.hardware [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 752.507852] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df02f2f1-eacd-499e-88b2-38de09b294d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.517212] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0bb46f-f334-49b2-bd55-0ce3809cc0bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.602042] env[65107]: DEBUG nova.scheduler.client.report [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 753.111183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.670s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.111433] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 753.114205] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.103s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.115641] env[65107]: INFO nova.compute.claims [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.412715] env[65107]: DEBUG nova.network.neutron [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Successfully updated port: 0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 753.463905] env[65107]: WARNING neutronclient.v2_0.client [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.464655] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.466847] env[65107]: WARNING openstack [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.607355] env[65107]: DEBUG nova.network.neutron [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.622962] env[65107]: DEBUG nova.compute.utils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 753.625265] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 753.625265] env[65107]: DEBUG nova.network.neutron [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 753.625626] env[65107]: WARNING neutronclient.v2_0.client [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.625977] env[65107]: WARNING neutronclient.v2_0.client [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.626711] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.627397] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.637809] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 753.685020] env[65107]: DEBUG nova.policy [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd87f09a43c049deb1fda21602b55a64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '544c649453844164bac53c98b342543f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 753.918752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.918752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquired lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.918996] env[65107]: DEBUG nova.network.neutron [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 754.068593] env[65107]: DEBUG nova.network.neutron [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Successfully created port: 8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 754.112021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.112021] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Instance network_info: |[{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 754.112384] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:5d:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3e0aae3-33d1-403b-bfaf-306f77a1422e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b44410c9-c28d-4e13-879d-3e495fde9ff4', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.119798] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Creating folder: Project (152868ee0723449baac10f6a8ceeb738). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.120173] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4362f78-1c08-4b25-aba6-3fa8427243fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.134438] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Created folder: Project (152868ee0723449baac10f6a8ceeb738) in parent group-v992574. [ 754.134666] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Creating folder: Instances. Parent ref: group-v992693. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.134919] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4756311f-1e7c-478c-9619-2a94d8ad9fe9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.152496] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Created folder: Instances in parent group-v992693. [ 754.153025] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.153672] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.153672] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a15b056-275f-4e05-997c-1883dffffc1d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.186525] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.186525] env[65107]: value = "task-5102622" [ 754.186525] env[65107]: _type = "Task" [ 754.186525] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.199893] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102622, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.422867] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.423178] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.466737] env[65107]: DEBUG nova.network.neutron [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 754.499951] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.500291] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.610019] env[65107]: WARNING neutronclient.v2_0.client [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.610019] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.610019] env[65107]: WARNING openstack [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.648552] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 754.678562] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 754.678832] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 754.678993] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 754.679778] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 754.680021] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 754.680183] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 754.680396] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.680555] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 754.680813] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 754.680994] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 754.681181] env[65107]: DEBUG nova.virt.hardware [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 754.682308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81363c6-4a80-4ee2-b916-ed71f1e785c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.693664] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50405ba-0ab3-4227-be65-9382d57c0c27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.703413] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102622, 'name': CreateVM_Task, 'duration_secs': 0.382052} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.704148] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 754.704754] env[65107]: WARNING neutronclient.v2_0.client [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.705138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.705288] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.705624] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 754.710143] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d51f747a-360c-4ea4-9296-a37111333661 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.727018] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 754.727018] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526ffa1f-4a0a-7f14-1153-6c5ff625277a" [ 754.727018] env[65107]: _type = "Task" [ 754.727018] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.735696] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526ffa1f-4a0a-7f14-1153-6c5ff625277a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.781754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2cb74e-9958-4d21-9dbe-8ce8dea8422f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.790544] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc46925a-6455-4c69-9435-78113f85c299 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.622102] env[65107]: DEBUG nova.network.neutron [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Updating instance_info_cache with network_info: [{"id": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "address": "fa:16:3e:e3:57:05", "network": {"id": "fa12197d-af9d-4a1b-8b2e-f5292d5da7a5", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1859092072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1fd7913632a4a5b9e076b5b156fa5fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d76af9c-1e", "ovs_interfaceid": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 755.625097] env[65107]: DEBUG nova.compute.manager [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-changed-b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 755.625305] env[65107]: DEBUG nova.compute.manager [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Refreshing instance network info cache due to event network-changed-b44410c9-c28d-4e13-879d-3e495fde9ff4. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 755.629029] env[65107]: DEBUG oslo_concurrency.lockutils [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Acquiring lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.629029] env[65107]: DEBUG oslo_concurrency.lockutils [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Acquired lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.629029] env[65107]: DEBUG nova.network.neutron [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Refreshing network info cache for port b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 755.629029] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e01422-0e3e-49ff-808c-c77efd1f05ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.641162] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f591da-7635-4f06-a00a-4838a680f93a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.645229] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526ffa1f-4a0a-7f14-1153-6c5ff625277a, 'name': SearchDatastore_Task, 'duration_secs': 0.036468} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.646339] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.646564] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.646811] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.647802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.647802] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.648321] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0273297a-57c3-46c5-a1b7-3df5926de331 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.659245] env[65107]: DEBUG nova.compute.provider_tree [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.670644] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.670834] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.673199] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebc435d6-b391-4b06-8c95-9a9dc3291e6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.681195] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 755.681195] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5296e5ae-40f0-67b3-4646-4eb59dc1a5c5" [ 755.681195] env[65107]: _type = "Task" [ 755.681195] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.689500] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5296e5ae-40f0-67b3-4646-4eb59dc1a5c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.766740] env[65107]: DEBUG nova.network.neutron [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Successfully updated port: 8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 755.843969] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.844334] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.136353] env[65107]: WARNING neutronclient.v2_0.client [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.136353] env[65107]: WARNING openstack [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.136353] env[65107]: WARNING openstack [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.142311] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Releasing lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.142665] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Instance network_info: |[{"id": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "address": "fa:16:3e:e3:57:05", "network": {"id": "fa12197d-af9d-4a1b-8b2e-f5292d5da7a5", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1859092072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1fd7913632a4a5b9e076b5b156fa5fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d76af9c-1e", "ovs_interfaceid": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 756.143387] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:57:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0d76af9c-1e77-4b0e-908b-551c74fe95ee', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 756.150866] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Creating folder: Project (e1fd7913632a4a5b9e076b5b156fa5fa). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 756.151175] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5908d08-e7c4-45ee-8c4a-cd1323f5c1c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.163136] env[65107]: DEBUG nova.scheduler.client.report [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 756.170073] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Created folder: Project (e1fd7913632a4a5b9e076b5b156fa5fa) in parent group-v992574. [ 756.172816] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Creating folder: Instances. Parent ref: group-v992696. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 756.173441] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0326bb0e-97ef-43e1-998f-1c3589239cd5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.188361] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Created folder: Instances in parent group-v992696. [ 756.188631] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 756.189901] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 756.190154] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f06933f-2895-4196-af19-29609e0c2760 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.209739] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5296e5ae-40f0-67b3-4646-4eb59dc1a5c5, 'name': SearchDatastore_Task, 'duration_secs': 0.023381} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.210635] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d91ef1c5-95dd-416a-98b2-811528afb9b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.216679] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 756.216679] env[65107]: value = "task-5102625" [ 756.216679] env[65107]: _type = "Task" [ 756.216679] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.218478] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 756.218478] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bcdf4c-d8f9-fb88-8dbc-3f73b725f3be" [ 756.218478] env[65107]: _type = "Task" [ 756.218478] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.234932] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bcdf4c-d8f9-fb88-8dbc-3f73b725f3be, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.235184] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102625, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.235449] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.235811] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] d2d0e18c-4839-416c-b363-b93996e8ba7f/d2d0e18c-4839-416c-b363-b93996e8ba7f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.236111] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3cf316c-1c13-4813-b533-05326b435456 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.243429] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 756.243429] env[65107]: value = "task-5102626" [ 756.243429] env[65107]: _type = "Task" [ 756.243429] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.251979] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.274356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.274356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.274663] env[65107]: DEBUG nova.network.neutron [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 756.347623] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 756.390481] env[65107]: WARNING openstack [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.391063] env[65107]: WARNING openstack [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.481485] env[65107]: DEBUG nova.compute.manager [req-1144e747-7b36-45c6-888e-848bb0a1f030 req-7692a700-5e12-488c-b0df-85add9492ff5 service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Received event network-vif-plugged-8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 756.481753] env[65107]: DEBUG oslo_concurrency.lockutils [req-1144e747-7b36-45c6-888e-848bb0a1f030 req-7692a700-5e12-488c-b0df-85add9492ff5 service nova] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.481939] env[65107]: DEBUG oslo_concurrency.lockutils [req-1144e747-7b36-45c6-888e-848bb0a1f030 req-7692a700-5e12-488c-b0df-85add9492ff5 service nova] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.482600] env[65107]: DEBUG oslo_concurrency.lockutils [req-1144e747-7b36-45c6-888e-848bb0a1f030 req-7692a700-5e12-488c-b0df-85add9492ff5 service nova] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.482795] env[65107]: DEBUG nova.compute.manager [req-1144e747-7b36-45c6-888e-848bb0a1f030 req-7692a700-5e12-488c-b0df-85add9492ff5 service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] No waiting events found dispatching network-vif-plugged-8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 756.482972] env[65107]: WARNING nova.compute.manager [req-1144e747-7b36-45c6-888e-848bb0a1f030 req-7692a700-5e12-488c-b0df-85add9492ff5 service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Received unexpected event network-vif-plugged-8622bbd6-6068-43cb-9da2-e996d2202ef2 for instance with vm_state building and task_state spawning. [ 756.484624] env[65107]: WARNING neutronclient.v2_0.client [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.486075] env[65107]: WARNING openstack [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.486075] env[65107]: WARNING openstack [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.584368] env[65107]: DEBUG nova.network.neutron [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updated VIF entry in instance network info cache for port b44410c9-c28d-4e13-879d-3e495fde9ff4. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 756.585270] env[65107]: DEBUG nova.network.neutron [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.674826] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.560s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.675557] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 756.683544] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.400s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.685554] env[65107]: INFO nova.compute.claims [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.731533] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102625, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.755381] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102626, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.778513] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.779548] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.852547] env[65107]: DEBUG nova.network.neutron [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 756.877577] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.967498] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.967900] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.088360] env[65107]: DEBUG oslo_concurrency.lockutils [req-e5cc9f18-4536-4ae5-b596-595dc2b55714 req-634afcba-96f1-4d21-a759-8a970bff0257 service nova] Releasing lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.113021] env[65107]: WARNING neutronclient.v2_0.client [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.113700] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.118398] env[65107]: WARNING openstack [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.195576] env[65107]: DEBUG nova.compute.utils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 757.197395] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 757.197395] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 757.198311] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.198311] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.198712] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.199076] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.234952] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102625, 'name': CreateVM_Task, 'duration_secs': 0.626319} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.235427] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.236773] env[65107]: WARNING neutronclient.v2_0.client [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.237680] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.237981] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.238440] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 757.238882] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8dedc77-573b-472c-8ade-8950d962b709 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.247728] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 757.247728] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c8d0d-dffe-b5d5-4917-a5cd41ccbaa2" [ 757.247728] env[65107]: _type = "Task" [ 757.247728] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.250726] env[65107]: DEBUG nova.network.neutron [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updating instance_info_cache with network_info: [{"id": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "address": "fa:16:3e:0b:13:8f", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8622bbd6-60", "ovs_interfaceid": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.271369] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591883} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.277019] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] d2d0e18c-4839-416c-b363-b93996e8ba7f/d2d0e18c-4839-416c-b363-b93996e8ba7f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 757.277019] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.277019] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c8d0d-dffe-b5d5-4917-a5cd41ccbaa2, 'name': SearchDatastore_Task, 'duration_secs': 0.011909} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.277019] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41a033e5-9fc1-4b01-bd15-27de3333bf7d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.278585] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.278941] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.280055] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.280055] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.280055] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.280419] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50bb9bc7-ec26-43a2-84de-3ec790d9fade {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.289794] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 757.289794] env[65107]: value = "task-5102627" [ 757.289794] env[65107]: _type = "Task" [ 757.289794] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.295047] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.295047] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.296066] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4de18726-686c-4baa-af27-3946a058bb67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.306344] env[65107]: DEBUG nova.policy [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73470fe64d63489a90389446368d1c51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1ef40d5301a473ea0a1691fecf01fd2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 757.312782] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102627, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.313494] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 757.313494] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e4e034-43d2-ea9b-7724-a0b1733cbdf3" [ 757.313494] env[65107]: _type = "Task" [ 757.313494] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.325041] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e4e034-43d2-ea9b-7724-a0b1733cbdf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.677905] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Successfully created port: 776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 757.707691] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 757.736354] env[65107]: DEBUG nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Received event network-vif-plugged-0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 757.739603] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Acquiring lock "bfe544ec-9b08-4118-a940-a51520ecaac0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.741055] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.004s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.741055] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.741055] env[65107]: DEBUG nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] No waiting events found dispatching network-vif-plugged-0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 757.741055] env[65107]: WARNING nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Received unexpected event network-vif-plugged-0d76af9c-1e77-4b0e-908b-551c74fe95ee for instance with vm_state building and task_state spawning. [ 757.741055] env[65107]: DEBUG nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Received event network-changed-0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 757.741258] env[65107]: DEBUG nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Refreshing instance network info cache due to event network-changed-0d76af9c-1e77-4b0e-908b-551c74fe95ee. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 757.741456] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Acquiring lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.742870] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Acquired lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.742870] env[65107]: DEBUG nova.network.neutron [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Refreshing network info cache for port 0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 757.761093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.761728] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Instance network_info: |[{"id": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "address": "fa:16:3e:0b:13:8f", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8622bbd6-60", "ovs_interfaceid": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 757.765908] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:13:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8622bbd6-6068-43cb-9da2-e996d2202ef2', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.777802] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Creating folder: Project (544c649453844164bac53c98b342543f). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.782212] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c4dc392-842f-45a9-ad3e-277456fd05a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.798537] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Created folder: Project (544c649453844164bac53c98b342543f) in parent group-v992574. [ 757.798784] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Creating folder: Instances. Parent ref: group-v992699. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.799401] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3981dda-fb58-44cd-b7dd-ac42b282a28f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.804638] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102627, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083459} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.804776] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.805611] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add377ae-c337-467e-be49-c469f277e200 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.817896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.818162] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.822845] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Created folder: Instances in parent group-v992699. [ 757.822845] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 757.835244] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.845543] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] d2d0e18c-4839-416c-b363-b93996e8ba7f/d2d0e18c-4839-416c-b363-b93996e8ba7f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.845891] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3dfd9864-875b-47ee-bafd-18c4d542706a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.861681] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-267382e8-c762-42c8-b98a-e72685dba9a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.886036] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e4e034-43d2-ea9b-7724-a0b1733cbdf3, 'name': SearchDatastore_Task, 'duration_secs': 0.023404} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.891239] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.891239] env[65107]: value = "task-5102630" [ 757.891239] env[65107]: _type = "Task" [ 757.891239] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.891239] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 757.891239] env[65107]: value = "task-5102631" [ 757.891239] env[65107]: _type = "Task" [ 757.891239] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.891239] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30a7daf5-dd7a-4b2a-9d3a-c4c9b79cc22b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.905155] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 757.905155] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a2ef74-b591-1e75-d836-4dbd74cc2b30" [ 757.905155] env[65107]: _type = "Task" [ 757.905155] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.910783] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102631, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.919073] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102630, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.927094] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a2ef74-b591-1e75-d836-4dbd74cc2b30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.140119] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Successfully created port: 5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 758.248523] env[65107]: WARNING neutronclient.v2_0.client [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.249227] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.250858] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.335038] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 758.405979] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102630, 'name': CreateVM_Task, 'duration_secs': 0.40184} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.412117] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 758.412523] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102631, 'name': ReconfigVM_Task, 'duration_secs': 0.423121} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.413483] env[65107]: WARNING neutronclient.v2_0.client [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.413858] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.413971] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.414268] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 758.414543] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Reconfigured VM instance instance-00000029 to attach disk [datastore1] d2d0e18c-4839-416c-b363-b93996e8ba7f/d2d0e18c-4839-416c-b363-b93996e8ba7f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.419283] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1099da5-1dd8-4eb9-bea6-cecadef10158 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.421204] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c72ea7d-1069-40e1-b733-2d7692d18fb4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.432407] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a2ef74-b591-1e75-d836-4dbd74cc2b30, 'name': SearchDatastore_Task, 'duration_secs': 0.021967} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.434830] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.435147] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] bfe544ec-9b08-4118-a940-a51520ecaac0/bfe544ec-9b08-4118-a940-a51520ecaac0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.435480] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 758.435480] env[65107]: value = "task-5102632" [ 758.435480] env[65107]: _type = "Task" [ 758.435480] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.435864] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 758.435864] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520603d2-f809-231b-8eea-e77f46113f0f" [ 758.435864] env[65107]: _type = "Task" [ 758.435864] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.436100] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89c8f75e-a856-43fe-9e81-16e68847f7f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.457832] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102632, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.462605] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520603d2-f809-231b-8eea-e77f46113f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.012304} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.462971] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 758.462971] env[65107]: value = "task-5102633" [ 758.462971] env[65107]: _type = "Task" [ 758.462971] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.466126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.466261] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.466878] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.466878] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.466878] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.467585] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74669760-0d26-4b17-b042-3458ffb571d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.484916] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.485212] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.486110] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd20bd29-3da5-4363-b73c-faed17730064 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.492915] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 758.492915] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dfd565-328b-660e-1b2d-4fc869f806e5" [ 758.492915] env[65107]: _type = "Task" [ 758.492915] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.504084] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dfd565-328b-660e-1b2d-4fc869f806e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.505248] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Successfully created port: 8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 758.509505] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0998387-f2ad-4a73-ab35-79020e7abb45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.518597] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebfc727-f6f3-441f-a0df-3314d472f5b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.561520] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678c3a41-c0a8-490e-ac7f-fddbabb447d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.574802] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e766f270-5d56-43ff-bf43-5a837205beed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.590730] env[65107]: DEBUG nova.compute.provider_tree [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.614236] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.614882] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.709490] env[65107]: WARNING neutronclient.v2_0.client [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.710168] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.710523] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.725448] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 758.769015] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 758.769321] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 758.769497] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 758.769708] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 758.769948] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 758.770280] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 758.770705] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.771061] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 758.771403] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 758.771662] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 758.772047] env[65107]: DEBUG nova.virt.hardware [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 758.773459] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db704a1a-2346-4a64-ad67-ab32bfc703d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.783930] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1addafae-6f03-4c77-b1f0-654f54aa4e87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.819369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.819639] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.823028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.823028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.823028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.823343] env[65107]: INFO nova.compute.manager [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Terminating instance [ 758.869056] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.902888] env[65107]: DEBUG nova.network.neutron [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Updated VIF entry in instance network info cache for port 0d76af9c-1e77-4b0e-908b-551c74fe95ee. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 758.903336] env[65107]: DEBUG nova.network.neutron [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Updating instance_info_cache with network_info: [{"id": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "address": "fa:16:3e:e3:57:05", "network": {"id": "fa12197d-af9d-4a1b-8b2e-f5292d5da7a5", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1859092072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1fd7913632a4a5b9e076b5b156fa5fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d76af9c-1e", "ovs_interfaceid": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.954510] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102632, 'name': Rename_Task, 'duration_secs': 0.195028} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.954510] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.954799] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25418f51-bfbe-4a15-a93a-46417bff349a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.964499] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 758.964499] env[65107]: value = "task-5102634" [ 758.964499] env[65107]: _type = "Task" [ 758.964499] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.979225] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102633, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.983333] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102634, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.991320] env[65107]: DEBUG nova.compute.manager [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Received event network-changed-8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 758.991320] env[65107]: DEBUG nova.compute.manager [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Refreshing instance network info cache due to event network-changed-8622bbd6-6068-43cb-9da2-e996d2202ef2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 758.991320] env[65107]: DEBUG oslo_concurrency.lockutils [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Acquiring lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.991320] env[65107]: DEBUG oslo_concurrency.lockutils [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Acquired lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.991320] env[65107]: DEBUG nova.network.neutron [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Refreshing network info cache for port 8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 759.007942] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dfd565-328b-660e-1b2d-4fc869f806e5, 'name': SearchDatastore_Task, 'duration_secs': 0.012849} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.009284] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24dfdfa1-df95-4ee3-8998-a0cea47d8305 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.019921] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 759.019921] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5258cfd5-1b74-547c-dd92-a445c510a53e" [ 759.019921] env[65107]: _type = "Task" [ 759.019921] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.033759] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5258cfd5-1b74-547c-dd92-a445c510a53e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.094114] env[65107]: DEBUG nova.scheduler.client.report [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 759.330181] env[65107]: DEBUG nova.compute.manager [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 759.330181] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.330817] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1153c4-5978-4072-9dfb-c3e5094fd863 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.339629] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 759.340336] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a7ba6b9-62df-4a67-865e-091bd0733550 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.349519] env[65107]: DEBUG oslo_vmware.api [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 759.349519] env[65107]: value = "task-5102635" [ 759.349519] env[65107]: _type = "Task" [ 759.349519] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.360516] env[65107]: DEBUG oslo_vmware.api [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.408034] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Releasing lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.408034] env[65107]: DEBUG nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Received event network-changed-33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 759.408034] env[65107]: DEBUG nova.compute.manager [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Refreshing instance network info cache due to event network-changed-33799b3d-5ade-4cb0-ae83-3d00bfd972ca. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 759.408034] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Acquiring lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.408034] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Acquired lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.408369] env[65107]: DEBUG nova.network.neutron [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Refreshing network info cache for port 33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 759.479133] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102633, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.650857} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.482879] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] bfe544ec-9b08-4118-a940-a51520ecaac0/bfe544ec-9b08-4118-a940-a51520ecaac0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 759.482879] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.483255] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102634, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.483435] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ca3096c-9a89-45bc-9406-f13ea518ed5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.491440] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 759.491440] env[65107]: value = "task-5102636" [ 759.491440] env[65107]: _type = "Task" [ 759.491440] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.495848] env[65107]: WARNING neutronclient.v2_0.client [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.498570] env[65107]: WARNING openstack [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.499204] env[65107]: WARNING openstack [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.514643] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.531323] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5258cfd5-1b74-547c-dd92-a445c510a53e, 'name': SearchDatastore_Task, 'duration_secs': 0.055862} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.531644] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.531896] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 82e8c093-b718-4d38-9682-ba2e710e4b93/82e8c093-b718-4d38-9682-ba2e710e4b93.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 759.532206] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-183fff2c-16da-4d91-8952-9be8323aef99 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.540285] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 759.540285] env[65107]: value = "task-5102637" [ 759.540285] env[65107]: _type = "Task" [ 759.540285] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.550660] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.603536] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.924s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.604418] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 759.608804] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.895s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.609203] env[65107]: DEBUG nova.objects.instance [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lazy-loading 'resources' on Instance uuid f7446d76-45c2-4e8b-981d-d37c230cf125 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 759.695465] env[65107]: WARNING openstack [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.696403] env[65107]: WARNING openstack [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.835237] env[65107]: WARNING neutronclient.v2_0.client [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.835237] env[65107]: WARNING openstack [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.836286] env[65107]: WARNING openstack [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.866755] env[65107]: DEBUG oslo_vmware.api [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102635, 'name': PowerOffVM_Task, 'duration_secs': 0.387488} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.867589] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 759.868108] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 759.868840] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87aabf65-7762-409e-8a4e-25e819d57799 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.911984] env[65107]: WARNING neutronclient.v2_0.client [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.913062] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.913704] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.953306] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 759.954239] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 759.954239] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Deleting the datastore file [datastore1] 0e6c9f2d-4412-4d8d-8348-6552a34ab349 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.954239] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ae993fd-dc4a-4b0e-bf94-a8b63080ec2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.964825] env[65107]: DEBUG oslo_vmware.api [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for the task: (returnval){ [ 759.964825] env[65107]: value = "task-5102639" [ 759.964825] env[65107]: _type = "Task" [ 759.964825] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.982289] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102634, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.986267] env[65107]: DEBUG oslo_vmware.api [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102639, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.007545] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.203822} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.007890] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 760.008833] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c34909f-a6da-454f-8b30-e111d0c39e6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.032817] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] bfe544ec-9b08-4118-a940-a51520ecaac0/bfe544ec-9b08-4118-a940-a51520ecaac0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.033593] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bedbf6c8-898c-4139-b69a-ba83039a23d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.054110] env[65107]: DEBUG nova.network.neutron [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updated VIF entry in instance network info cache for port 8622bbd6-6068-43cb-9da2-e996d2202ef2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 760.054608] env[65107]: DEBUG nova.network.neutron [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updating instance_info_cache with network_info: [{"id": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "address": "fa:16:3e:0b:13:8f", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8622bbd6-60", "ovs_interfaceid": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 760.071677] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102637, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.074842] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 760.074842] env[65107]: value = "task-5102640" [ 760.074842] env[65107]: _type = "Task" [ 760.074842] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.092459] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102640, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.113387] env[65107]: DEBUG nova.compute.utils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 760.118155] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 760.118346] env[65107]: DEBUG nova.network.neutron [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 760.118690] env[65107]: WARNING neutronclient.v2_0.client [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.118977] env[65107]: WARNING neutronclient.v2_0.client [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.119745] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.119942] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.171559] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.171977] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.261533] env[65107]: DEBUG nova.policy [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32155078a41249a297470fd4083860df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f9188f3e35e4d06a977e2180918c616', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 760.297312] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Successfully updated port: 776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 760.485504] env[65107]: DEBUG oslo_vmware.api [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102634, 'name': PowerOnVM_Task, 'duration_secs': 1.041834} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.485746] env[65107]: DEBUG oslo_vmware.api [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Task: {'id': task-5102639, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.512299} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.486020] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 760.486236] env[65107]: INFO nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Took 10.74 seconds to spawn the instance on the hypervisor. [ 760.486474] env[65107]: DEBUG nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 760.486688] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.486855] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 760.487045] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.487242] env[65107]: INFO nova.compute.manager [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Took 1.16 seconds to destroy the instance on the hypervisor. [ 760.487446] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 760.490985] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94879a94-9049-4d36-becb-b0e4303400d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.493791] env[65107]: DEBUG nova.compute.manager [-] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 760.493930] env[65107]: DEBUG nova.network.neutron [-] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 760.494187] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.494707] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.494961] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.503862] env[65107]: WARNING neutronclient.v2_0.client [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.504541] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.504880] env[65107]: WARNING openstack [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.561146] env[65107]: DEBUG oslo_concurrency.lockutils [req-b5ee46b0-d99d-4e14-bd58-d4f38a9f3421 req-839167fa-1b1b-4c50-bb71-e890458ac78f service nova] Releasing lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.571970] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.872405} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.572796] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 82e8c093-b718-4d38-9682-ba2e710e4b93/82e8c093-b718-4d38-9682-ba2e710e4b93.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 760.572796] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 760.574093] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f56d8b3a-0dc2-4aa0-9b34-7c0b0075ceed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.588495] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 760.588495] env[65107]: value = "task-5102641" [ 760.588495] env[65107]: _type = "Task" [ 760.588495] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.592976] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102640, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.604466] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102641, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.607145] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.629276] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 760.737018] env[65107]: DEBUG nova.network.neutron [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Successfully created port: a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 760.808520] env[65107]: DEBUG nova.network.neutron [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updated VIF entry in instance network info cache for port 33799b3d-5ade-4cb0-ae83-3d00bfd972ca. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 760.808886] env[65107]: DEBUG nova.network.neutron [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updating instance_info_cache with network_info: [{"id": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "address": "fa:16:3e:27:e4:ae", "network": {"id": "472c4d13-7d57-4d84-b35d-7f05f937120a", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1598181546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71cd673a320348a9925e8f59637afc17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33799b3d-5a", "ovs_interfaceid": "33799b3d-5ade-4cb0-ae83-3d00bfd972ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 760.831583] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9f368a-3a91-4f04-966e-6bc2569b0127 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.842186] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d4e842-772f-45e0-b400-da51f6ceef9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.878656] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84773e23-7c39-4730-ad19-da6c6dd4d41b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.887540] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b363261-3c72-45db-b9eb-cf2b4227bcc1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.903585] env[65107]: DEBUG nova.compute.provider_tree [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.033029] env[65107]: INFO nova.compute.manager [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Took 38.22 seconds to build instance. [ 761.091274] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102640, 'name': ReconfigVM_Task, 'duration_secs': 0.596142} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.091824] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Reconfigured VM instance instance-0000002a to attach disk [datastore1] bfe544ec-9b08-4118-a940-a51520ecaac0/bfe544ec-9b08-4118-a940-a51520ecaac0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.092820] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d449061-95b2-46e2-8dc4-4de1e39f3d60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.107082] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102641, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.163362} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.109100] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.109597] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 761.109597] env[65107]: value = "task-5102642" [ 761.109597] env[65107]: _type = "Task" [ 761.109597] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.110907] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351c2375-9e6d-4af9-b4ad-60678040af1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.153943] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 82e8c093-b718-4d38-9682-ba2e710e4b93/82e8c093-b718-4d38-9682-ba2e710e4b93.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.158488] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2598fa0-73c0-48e9-98af-4c58fd5fc7a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.174023] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102642, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.180934] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 761.180934] env[65107]: value = "task-5102643" [ 761.180934] env[65107]: _type = "Task" [ 761.180934] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.192390] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102643, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.314361] env[65107]: DEBUG oslo_concurrency.lockutils [req-5d842a2c-409a-4036-b284-f57718cb3761 req-2b2d83df-d6ee-4731-a5d8-deec9983dc3f service nova] Releasing lock "refresh_cache-d65e98bb-f54d-4803-95ca-0fc9b6c184bc" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.406961] env[65107]: DEBUG nova.scheduler.client.report [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.535647] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6b357707-17ff-481b-ab77-6d39da2a0efc tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.086s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.626321] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102642, 'name': Rename_Task, 'duration_secs': 0.169577} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.626590] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.626921] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d83cc90a-3a2e-4a11-a307-de00cf647869 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.633545] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 761.633545] env[65107]: value = "task-5102644" [ 761.633545] env[65107]: _type = "Task" [ 761.633545] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.641725] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.656166] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 761.684320] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 761.684581] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.684735] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 761.684917] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.685075] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 761.685226] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 761.685433] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.685588] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 761.685749] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 761.685946] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 761.686157] env[65107]: DEBUG nova.virt.hardware [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 761.687149] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaefc75-5b3b-4968-806d-8c85ed8d0fb4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.699947] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102643, 'name': ReconfigVM_Task, 'duration_secs': 0.357068} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.702493] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 82e8c093-b718-4d38-9682-ba2e710e4b93/82e8c093-b718-4d38-9682-ba2e710e4b93.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.703176] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a09423c-aa80-45f1-aecb-0d01cd3cba50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.706209] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cc043b-ee44-4113-b476-3491aba0e2f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.722637] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 761.722637] env[65107]: value = "task-5102645" [ 761.722637] env[65107]: _type = "Task" [ 761.722637] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.731618] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102645, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.831274] env[65107]: DEBUG nova.network.neutron [-] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 761.913747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.305s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.916506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.429s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.918260] env[65107]: INFO nova.compute.claims [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.955195] env[65107]: INFO nova.scheduler.client.report [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Deleted allocations for instance f7446d76-45c2-4e8b-981d-d37c230cf125 [ 762.143782] env[65107]: DEBUG oslo_vmware.api [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5102644, 'name': PowerOnVM_Task, 'duration_secs': 0.493192} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.144065] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.144284] env[65107]: INFO nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Took 9.66 seconds to spawn the instance on the hypervisor. [ 762.144497] env[65107]: DEBUG nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 762.145295] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4da2b4-e781-4ef6-a6fb-da309a0920ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.233275] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102645, 'name': Rename_Task, 'duration_secs': 0.151304} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.233558] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.233803] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-629e6149-c5f9-4243-aadc-d21168045c13 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.241561] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 762.241561] env[65107]: value = "task-5102646" [ 762.241561] env[65107]: _type = "Task" [ 762.241561] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.250626] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.334461] env[65107]: INFO nova.compute.manager [-] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Took 1.84 seconds to deallocate network for instance. [ 762.463843] env[65107]: DEBUG oslo_concurrency.lockutils [None req-041568a7-927f-4cdd-a903-15b97e7c618c tempest-ImagesOneServerTestJSON-2092730972 tempest-ImagesOneServerTestJSON-2092730972-project-member] Lock "f7446d76-45c2-4e8b-981d-d37c230cf125" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.494s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.531082] env[65107]: DEBUG nova.network.neutron [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Successfully updated port: a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 762.665931] env[65107]: INFO nova.compute.manager [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Took 34.90 seconds to build instance. [ 762.754304] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102646, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.842171] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.913819] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Successfully updated port: 5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 763.037388] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "refresh_cache-19518d90-ba6f-44e5-b7c7-1ececad2b54d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.037596] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "refresh_cache-19518d90-ba6f-44e5-b7c7-1ececad2b54d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.037771] env[65107]: DEBUG nova.network.neutron [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 763.119320] env[65107]: DEBUG nova.compute.manager [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-vif-plugged-776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 763.119647] env[65107]: DEBUG oslo_concurrency.lockutils [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Acquiring lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.119816] env[65107]: DEBUG oslo_concurrency.lockutils [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.119973] env[65107]: DEBUG oslo_concurrency.lockutils [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.120146] env[65107]: DEBUG nova.compute.manager [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] No waiting events found dispatching network-vif-plugged-776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 763.120331] env[65107]: WARNING nova.compute.manager [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received unexpected event network-vif-plugged-776f5e6b-b550-48cb-9141-89cfb94a6460 for instance with vm_state building and task_state spawning. [ 763.120493] env[65107]: DEBUG nova.compute.manager [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-changed-776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 763.120642] env[65107]: DEBUG nova.compute.manager [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Refreshing instance network info cache due to event network-changed-776f5e6b-b550-48cb-9141-89cfb94a6460. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 763.120830] env[65107]: DEBUG oslo_concurrency.lockutils [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Acquiring lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.121096] env[65107]: DEBUG oslo_concurrency.lockutils [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Acquired lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.121158] env[65107]: DEBUG nova.network.neutron [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Refreshing network info cache for port 776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 763.168717] env[65107]: DEBUG oslo_concurrency.lockutils [None req-002003e7-ad1e-49d0-97df-908ef0762553 tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.518s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.254852] env[65107]: DEBUG oslo_vmware.api [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102646, 'name': PowerOnVM_Task, 'duration_secs': 0.580836} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.255131] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.255332] env[65107]: INFO nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Took 8.61 seconds to spawn the instance on the hypervisor. [ 763.255509] env[65107]: DEBUG nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 763.256356] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb5428b-fe86-4638-bd03-ec9560896425 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.402692] env[65107]: DEBUG nova.compute.manager [req-b9b1534d-b8d9-4ef8-90ff-eed5d87b54aa req-f2bf7b8c-b381-4118-91bd-a20df568de11 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Received event network-vif-plugged-a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 763.402926] env[65107]: DEBUG oslo_concurrency.lockutils [req-b9b1534d-b8d9-4ef8-90ff-eed5d87b54aa req-f2bf7b8c-b381-4118-91bd-a20df568de11 service nova] Acquiring lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.403154] env[65107]: DEBUG oslo_concurrency.lockutils [req-b9b1534d-b8d9-4ef8-90ff-eed5d87b54aa req-f2bf7b8c-b381-4118-91bd-a20df568de11 service nova] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.403324] env[65107]: DEBUG oslo_concurrency.lockutils [req-b9b1534d-b8d9-4ef8-90ff-eed5d87b54aa req-f2bf7b8c-b381-4118-91bd-a20df568de11 service nova] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.403490] env[65107]: DEBUG nova.compute.manager [req-b9b1534d-b8d9-4ef8-90ff-eed5d87b54aa req-f2bf7b8c-b381-4118-91bd-a20df568de11 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] No waiting events found dispatching network-vif-plugged-a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 763.403693] env[65107]: WARNING nova.compute.manager [req-b9b1534d-b8d9-4ef8-90ff-eed5d87b54aa req-f2bf7b8c-b381-4118-91bd-a20df568de11 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Received unexpected event network-vif-plugged-a6da0470-a2c9-4d2d-869b-c102c75b6b54 for instance with vm_state building and task_state spawning. [ 763.452401] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b53526e-176f-413b-8c88-1c49f5588be9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.461221] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254435dc-6021-4338-be72-03224a43b195 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.496133] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc039ea3-97e2-4e05-a7ce-a92a8d284c4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.504854] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4aa177-e381-4e19-b5a5-f75ce8769e41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.521795] env[65107]: DEBUG nova.compute.provider_tree [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.541730] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.542241] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.626498] env[65107]: WARNING neutronclient.v2_0.client [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.627249] env[65107]: WARNING openstack [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.627558] env[65107]: WARNING openstack [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.780994] env[65107]: INFO nova.compute.manager [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Took 35.96 seconds to build instance. [ 763.894396] env[65107]: DEBUG nova.network.neutron [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 763.928092] env[65107]: DEBUG nova.network.neutron [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 764.024969] env[65107]: DEBUG nova.scheduler.client.report [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.282776] env[65107]: DEBUG oslo_concurrency.lockutils [None req-061aff56-a2c3-489a-b2be-f1dc54dea521 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.121s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.530628] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.530995] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 764.536875] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.025s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.536875] env[65107]: DEBUG nova.objects.instance [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 764.981171] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Successfully updated port: 8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 765.043738] env[65107]: DEBUG nova.compute.utils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 765.048946] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.049631] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.062265] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 765.062986] env[65107]: DEBUG nova.network.neutron [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 765.062986] env[65107]: WARNING neutronclient.v2_0.client [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.063449] env[65107]: WARNING neutronclient.v2_0.client [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.064322] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.064873] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.325837] env[65107]: DEBUG nova.policy [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75dbe2746e84cc2ad6848d6bd5412b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58588fe2a11f4ee0aab313903b211fa8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 765.390846] env[65107]: DEBUG nova.network.neutron [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 765.487031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.512643] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.512643] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.512643] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.512643] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.512643] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.512643] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.513222] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.513222] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 765.513222] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.544396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2b776a4-3aec-43e7-b025-27f441b1d01a tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.546278] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.373s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.546721] env[65107]: DEBUG nova.objects.instance [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lazy-loading 'resources' on Instance uuid 705b91ea-ccd2-41ce-a14f-2ea5b70f80df {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.548421] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 765.679628] env[65107]: WARNING neutronclient.v2_0.client [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.679628] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.679628] env[65107]: WARNING openstack [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.897104] env[65107]: DEBUG oslo_concurrency.lockutils [req-ae1a5812-570a-4720-9ba6-47bf28486cfe req-3320ec3d-7171-46c0-82dd-c2bad47efc9f service nova] Releasing lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.897687] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.897935] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 765.999963] env[65107]: DEBUG nova.network.neutron [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Successfully created port: 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 766.016450] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.050316] env[65107]: DEBUG nova.network.neutron [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Updating instance_info_cache with network_info: [{"id": "a6da0470-a2c9-4d2d-869b-c102c75b6b54", "address": "fa:16:3e:f9:4b:a2", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6da0470-a2", "ovs_interfaceid": "a6da0470-a2c9-4d2d-869b-c102c75b6b54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.401574] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.401985] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.458793] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 766.508367] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.508795] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.557943] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "refresh_cache-19518d90-ba6f-44e5-b7c7-1ececad2b54d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.558418] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Instance network_info: |[{"id": "a6da0470-a2c9-4d2d-869b-c102c75b6b54", "address": "fa:16:3e:f9:4b:a2", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6da0470-a2", "ovs_interfaceid": "a6da0470-a2c9-4d2d-869b-c102c75b6b54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 766.561486] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:4b:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6da0470-a2c9-4d2d-869b-c102c75b6b54', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 766.569177] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating folder: Project (6f9188f3e35e4d06a977e2180918c616). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.570555] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 766.572544] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-707fa230-4af6-450f-b8e3-23845437671a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.588400] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created folder: Project (6f9188f3e35e4d06a977e2180918c616) in parent group-v992574. [ 766.588400] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating folder: Instances. Parent ref: group-v992702. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.588400] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c31735e8-877d-45ee-ac87-bb49d2dd94ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.596640] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.597585] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.597889] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.610879] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created folder: Instances in parent group-v992702. [ 766.611165] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 766.611966] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 766.611966] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0dc8b494-9a70-491d-b8df-b520baa15daa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.631089] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 766.631337] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.631475] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 766.631653] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.631794] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 766.631937] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 766.632157] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.632314] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 766.632473] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 766.632630] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 766.632800] env[65107]: DEBUG nova.virt.hardware [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 766.633685] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bf002a-cdf9-4aac-b79e-2d854fd3fe0c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.643417] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbdcbaa-dfdb-4cb7-b444-2a3884a44b1e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.648519] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 766.648519] env[65107]: value = "task-5102649" [ 766.648519] env[65107]: _type = "Task" [ 766.648519] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.669151] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102649, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.718719] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3edaa03-4a01-4af2-86ac-5f4f55378fba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.726718] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57285503-7da4-49d5-8f2d-fa6b79371e03 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.758728] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee58c628-f32c-4190-bc4c-438f92c05b8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.768592] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c1a85e-f897-438c-badd-73463d25477e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.782980] env[65107]: DEBUG nova.compute.provider_tree [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.833639] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.834099] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.851176] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Received event network-vif-deleted-55102692-2502-4633-b00b-cc29760fb310 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 766.851381] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-vif-plugged-5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 766.851613] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Acquiring lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.851767] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.851934] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.852118] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] No waiting events found dispatching network-vif-plugged-5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 766.852282] env[65107]: WARNING nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received unexpected event network-vif-plugged-5a370d18-6b6f-48cb-b7ea-50951741486f for instance with vm_state building and task_state spawning. [ 766.852436] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-changed-5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 766.853023] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Refreshing instance network info cache due to event network-changed-5a370d18-6b6f-48cb-b7ea-50951741486f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 766.853023] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Acquiring lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.875489] env[65107]: DEBUG nova.compute.manager [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Received event network-changed-a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 766.875712] env[65107]: DEBUG nova.compute.manager [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Refreshing instance network info cache due to event network-changed-a6da0470-a2c9-4d2d-869b-c102c75b6b54. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 766.875970] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Acquiring lock "refresh_cache-19518d90-ba6f-44e5-b7c7-1ececad2b54d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.876394] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Acquired lock "refresh_cache-19518d90-ba6f-44e5-b7c7-1ececad2b54d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.876603] env[65107]: DEBUG nova.network.neutron [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Refreshing network info cache for port a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 766.900988] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.901709] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.902212] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.162468] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102649, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.191937] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.192529] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.286460] env[65107]: DEBUG nova.scheduler.client.report [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 767.350668] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.351429] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.352100] env[65107]: WARNING openstack [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.378943] env[65107]: WARNING neutronclient.v2_0.client [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.379791] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.380156] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.667783] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102649, 'name': CreateVM_Task, 'duration_secs': 0.57695} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.671024] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 767.671024] env[65107]: WARNING neutronclient.v2_0.client [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.672504] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.672674] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.673030] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 767.674093] env[65107]: DEBUG nova.network.neutron [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Successfully updated port: 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 767.675508] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c5e4a93-3239-4d34-bccd-a937b21cc58e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.684891] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 767.684891] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529072cb-48a5-a8de-2d8e-8efbc358846e" [ 767.684891] env[65107]: _type = "Task" [ 767.684891] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.700530] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529072cb-48a5-a8de-2d8e-8efbc358846e, 'name': SearchDatastore_Task, 'duration_secs': 0.011887} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.701363] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.701363] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 767.701562] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.701745] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.702184] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.702254] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac00fa5c-f24b-4de6-a8af-c5e07c318830 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.713141] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.713473] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 767.714989] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7cacaac-3bfe-44bf-887d-127231b93311 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.721832] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 767.721832] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e7f5e6-778b-1efe-7bda-bbbb7842e2c2" [ 767.721832] env[65107]: _type = "Task" [ 767.721832] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.735581] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e7f5e6-778b-1efe-7bda-bbbb7842e2c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.793406] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.247s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.796778] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.286s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.796778] env[65107]: DEBUG nova.objects.instance [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lazy-loading 'resources' on Instance uuid 3948dcdd-e7c1-40b6-a3ec-97741940a0fb {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.809189] env[65107]: DEBUG nova.network.neutron [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [{"id": "776f5e6b-b550-48cb-9141-89cfb94a6460", "address": "fa:16:3e:d2:f3:f6", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap776f5e6b-b5", "ovs_interfaceid": "776f5e6b-b550-48cb-9141-89cfb94a6460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5a370d18-6b6f-48cb-b7ea-50951741486f", "address": "fa:16:3e:5d:29:47", "network": {"id": "c771b8d1-00f5-4f8d-93a6-5635a74f7798", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1610194998", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.114", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a370d18-6b", "ovs_interfaceid": "5a370d18-6b6f-48cb-b7ea-50951741486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "address": "fa:16:3e:f4:3b:4f", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5dd2fd-4c", "ovs_interfaceid": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 767.837173] env[65107]: INFO nova.scheduler.client.report [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Deleted allocations for instance 705b91ea-ccd2-41ce-a14f-2ea5b70f80df [ 767.940138] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.940539] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.044388] env[65107]: WARNING neutronclient.v2_0.client [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.045166] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.045524] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.160624] env[65107]: DEBUG nova.network.neutron [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Updated VIF entry in instance network info cache for port a6da0470-a2c9-4d2d-869b-c102c75b6b54. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 768.161031] env[65107]: DEBUG nova.network.neutron [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Updating instance_info_cache with network_info: [{"id": "a6da0470-a2c9-4d2d-869b-c102c75b6b54", "address": "fa:16:3e:f9:4b:a2", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6da0470-a2", "ovs_interfaceid": "a6da0470-a2c9-4d2d-869b-c102c75b6b54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 768.180781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.181123] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.181332] env[65107]: DEBUG nova.network.neutron [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 768.194043] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "780f910f-4f73-41ba-a795-0daae3097314" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.194043] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "780f910f-4f73-41ba-a795-0daae3097314" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.239645] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e7f5e6-778b-1efe-7bda-bbbb7842e2c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010791} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.240369] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf8210f7-f9e0-450b-8e5e-ed12c37ca5ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.246495] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 768.246495] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1b588-14bd-4847-f3d6-0bedd5d22d07" [ 768.246495] env[65107]: _type = "Task" [ 768.246495] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.256996] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1b588-14bd-4847-f3d6-0bedd5d22d07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.311593] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Releasing lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.312479] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Instance network_info: |[{"id": "776f5e6b-b550-48cb-9141-89cfb94a6460", "address": "fa:16:3e:d2:f3:f6", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap776f5e6b-b5", "ovs_interfaceid": "776f5e6b-b550-48cb-9141-89cfb94a6460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5a370d18-6b6f-48cb-b7ea-50951741486f", "address": "fa:16:3e:5d:29:47", "network": {"id": "c771b8d1-00f5-4f8d-93a6-5635a74f7798", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1610194998", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.114", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a370d18-6b", "ovs_interfaceid": "5a370d18-6b6f-48cb-b7ea-50951741486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "address": "fa:16:3e:f4:3b:4f", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5dd2fd-4c", "ovs_interfaceid": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 768.312479] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Acquired lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.312688] env[65107]: DEBUG nova.network.neutron [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Refreshing network info cache for port 5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 768.314469] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:f3:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1b8b991-feba-44e6-900c-6486e7e122f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '776f5e6b-b550-48cb-9141-89cfb94a6460', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:29:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57691231-2b8d-4d71-8f79-d4a6a1d95ec8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a370d18-6b6f-48cb-b7ea-50951741486f', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:3b:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1b8b991-feba-44e6-900c-6486e7e122f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.330301] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Creating folder: Project (d1ef40d5301a473ea0a1691fecf01fd2). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 768.332364] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffcd0b53-b950-4bb7-a0c9-2def425969f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.345513] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c339fa5-3483-4871-a366-f6acf4e5b619 tempest-ServersTestJSON-86747762 tempest-ServersTestJSON-86747762-project-member] Lock "705b91ea-ccd2-41ce-a14f-2ea5b70f80df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.182s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.348974] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Created folder: Project (d1ef40d5301a473ea0a1691fecf01fd2) in parent group-v992574. [ 768.349188] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Creating folder: Instances. Parent ref: group-v992705. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 768.349443] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8af1ff6-fda9-4630-85f2-f95518e91df9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.364426] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Created folder: Instances in parent group-v992705. [ 768.364611] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 768.365716] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 768.366169] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f98237e-2b01-4220-a9ea-a7f3f947324e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.395086] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.395086] env[65107]: value = "task-5102652" [ 768.395086] env[65107]: _type = "Task" [ 768.395086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.407280] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102652, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.664134] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Releasing lock "refresh_cache-19518d90-ba6f-44e5-b7c7-1ececad2b54d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.664466] env[65107]: DEBUG nova.compute.manager [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-changed-b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 768.664637] env[65107]: DEBUG nova.compute.manager [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Refreshing instance network info cache due to event network-changed-b44410c9-c28d-4e13-879d-3e495fde9ff4. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 768.664951] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Acquiring lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.665037] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Acquired lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.665259] env[65107]: DEBUG nova.network.neutron [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Refreshing network info cache for port b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 768.685190] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.685605] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.696830] env[65107]: DEBUG nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 768.731645] env[65107]: DEBUG nova.network.neutron [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 768.755146] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.755583] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.769429] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1b588-14bd-4847-f3d6-0bedd5d22d07, 'name': SearchDatastore_Task, 'duration_secs': 0.027063} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.769766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.770045] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 19518d90-ba6f-44e5-b7c7-1ececad2b54d/19518d90-ba6f-44e5-b7c7-1ececad2b54d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.770320] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9930f398-9bfa-491d-b1a4-fe87f7fd8ce6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.781264] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 768.781264] env[65107]: value = "task-5102653" [ 768.781264] env[65107]: _type = "Task" [ 768.781264] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.800671] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102653, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.823680] env[65107]: WARNING neutronclient.v2_0.client [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.824748] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.824985] env[65107]: WARNING openstack [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.833407] env[65107]: WARNING neutronclient.v2_0.client [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.834034] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.834423] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.885345] env[65107]: DEBUG nova.compute.manager [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Received event network-vif-plugged-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 768.885345] env[65107]: DEBUG oslo_concurrency.lockutils [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Acquiring lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.885850] env[65107]: DEBUG oslo_concurrency.lockutils [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.886245] env[65107]: DEBUG oslo_concurrency.lockutils [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.886616] env[65107]: DEBUG nova.compute.manager [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] No waiting events found dispatching network-vif-plugged-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 768.886902] env[65107]: WARNING nova.compute.manager [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Received unexpected event network-vif-plugged-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 for instance with vm_state building and task_state spawning. [ 768.887291] env[65107]: DEBUG nova.compute.manager [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Received event network-changed-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 768.887581] env[65107]: DEBUG nova.compute.manager [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Refreshing instance network info cache due to event network-changed-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 768.888135] env[65107]: DEBUG oslo_concurrency.lockutils [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Acquiring lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.910075] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102652, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.912586] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca995f74-6661-4334-a95b-1d2ee4b6649d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.921405] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976f9aba-2338-496d-b980-87271c7a11bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.928102] env[65107]: DEBUG nova.compute.manager [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Received event network-changed-0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 768.928102] env[65107]: DEBUG nova.compute.manager [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Refreshing instance network info cache due to event network-changed-0d76af9c-1e77-4b0e-908b-551c74fe95ee. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 768.928619] env[65107]: DEBUG oslo_concurrency.lockutils [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Acquiring lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.928619] env[65107]: DEBUG oslo_concurrency.lockutils [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Acquired lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.928619] env[65107]: DEBUG nova.network.neutron [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Refreshing network info cache for port 0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 768.970828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce3a85d-30a4-4699-873b-cae9832ebf68 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.980674] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f996c7-8ce0-4344-8c32-f7fe6304cb34 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.997137] env[65107]: DEBUG nova.compute.provider_tree [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.001927] env[65107]: DEBUG nova.network.neutron [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [{"id": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "address": "fa:16:3e:e0:45:0e", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap257c9dd4-8a", "ovs_interfaceid": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.134763] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.135178] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.171942] env[65107]: WARNING neutronclient.v2_0.client [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.172700] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.173185] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.222584] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.294386] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102653, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.300093] env[65107]: WARNING neutronclient.v2_0.client [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.301272] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.301781] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.411258] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102652, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.439030] env[65107]: WARNING neutronclient.v2_0.client [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.439030] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.439030] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.475764] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.476252] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.500932] env[65107]: DEBUG nova.scheduler.client.report [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 769.507779] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.508241] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance network_info: |[{"id": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "address": "fa:16:3e:e0:45:0e", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap257c9dd4-8a", "ovs_interfaceid": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 769.508940] env[65107]: DEBUG oslo_concurrency.lockutils [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Acquired lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.509145] env[65107]: DEBUG nova.network.neutron [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Refreshing network info cache for port 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 769.510327] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:45:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee4b2432-c393-4e50-ae0e-b5e12bad37db', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '257c9dd4-8ac7-4bca-abb6-a32d3652c6a2', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.517824] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 769.518434] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.518721] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbfd4f19-d9ef-41e3-92a8-870651229a4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.537491] env[65107]: DEBUG nova.network.neutron [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updated VIF entry in instance network info cache for port 5a370d18-6b6f-48cb-b7ea-50951741486f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 769.537978] env[65107]: DEBUG nova.network.neutron [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [{"id": "776f5e6b-b550-48cb-9141-89cfb94a6460", "address": "fa:16:3e:d2:f3:f6", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap776f5e6b-b5", "ovs_interfaceid": "776f5e6b-b550-48cb-9141-89cfb94a6460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5a370d18-6b6f-48cb-b7ea-50951741486f", "address": "fa:16:3e:5d:29:47", "network": {"id": "c771b8d1-00f5-4f8d-93a6-5635a74f7798", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1610194998", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.114", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a370d18-6b", "ovs_interfaceid": "5a370d18-6b6f-48cb-b7ea-50951741486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "address": "fa:16:3e:f4:3b:4f", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5dd2fd-4c", "ovs_interfaceid": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.545673] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.545673] env[65107]: value = "task-5102654" [ 769.545673] env[65107]: _type = "Task" [ 769.545673] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.557766] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102654, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.676419] env[65107]: WARNING neutronclient.v2_0.client [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.677194] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.677651] env[65107]: WARNING openstack [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.713954] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.714374] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.797763] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102653, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.716415} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.798083] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 19518d90-ba6f-44e5-b7c7-1ececad2b54d/19518d90-ba6f-44e5-b7c7-1ececad2b54d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.798329] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.801671] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53623251-1215-4a45-99d7-f6b0f2c975f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.808201] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 769.808201] env[65107]: value = "task-5102655" [ 769.808201] env[65107]: _type = "Task" [ 769.808201] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.820073] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102655, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.843324] env[65107]: DEBUG nova.network.neutron [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updated VIF entry in instance network info cache for port b44410c9-c28d-4e13-879d-3e495fde9ff4. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 769.843772] env[65107]: DEBUG nova.network.neutron [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.877321] env[65107]: WARNING neutronclient.v2_0.client [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.878025] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.878387] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.909458] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102652, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.011037] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.215s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.014270] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.231s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.014614] env[65107]: DEBUG nova.objects.instance [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lazy-loading 'resources' on Instance uuid b20a2c71-514d-488a-8edb-8fb0274a894a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 770.019784] env[65107]: WARNING neutronclient.v2_0.client [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.020566] env[65107]: WARNING openstack [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.020987] env[65107]: WARNING openstack [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.040540] env[65107]: INFO nova.scheduler.client.report [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleted allocations for instance 3948dcdd-e7c1-40b6-a3ec-97741940a0fb [ 770.043059] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Releasing lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.043059] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-vif-plugged-8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 770.044618] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Acquiring lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.044618] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.044618] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.044618] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] No waiting events found dispatching network-vif-plugged-8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 770.044618] env[65107]: WARNING nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received unexpected event network-vif-plugged-8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 for instance with vm_state building and task_state spawning. [ 770.044618] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-changed-8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 770.044618] env[65107]: DEBUG nova.compute.manager [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Refreshing instance network info cache due to event network-changed-8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 770.044618] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Acquiring lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.045805] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Acquired lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.045805] env[65107]: DEBUG nova.network.neutron [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Refreshing network info cache for port 8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 770.056149] env[65107]: DEBUG nova.network.neutron [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Updated VIF entry in instance network info cache for port 0d76af9c-1e77-4b0e-908b-551c74fe95ee. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 770.056573] env[65107]: DEBUG nova.network.neutron [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Updating instance_info_cache with network_info: [{"id": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "address": "fa:16:3e:e3:57:05", "network": {"id": "fa12197d-af9d-4a1b-8b2e-f5292d5da7a5", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1859092072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1fd7913632a4a5b9e076b5b156fa5fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d76af9c-1e", "ovs_interfaceid": "0d76af9c-1e77-4b0e-908b-551c74fe95ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.065749] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102654, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.220249] env[65107]: WARNING openstack [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.220396] env[65107]: WARNING openstack [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.325829] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102655, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07074} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.326655] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.326984] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cc9795-cdb3-4863-a4b2-25beaca5a4cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.351936] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 19518d90-ba6f-44e5-b7c7-1ececad2b54d/19518d90-ba6f-44e5-b7c7-1ececad2b54d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 770.353519] env[65107]: WARNING neutronclient.v2_0.client [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.354283] env[65107]: WARNING openstack [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.354634] env[65107]: WARNING openstack [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.362644] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a2e2383-d77f-4783-be16-5f672d8c59c6 req-b115a9f8-d0f6-444c-9690-8caab1e00e30 service nova] Releasing lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.363028] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fd359b3-92e6-4eba-b7dc-a5eaea7b2cc0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.394618] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 770.394618] env[65107]: value = "task-5102656" [ 770.394618] env[65107]: _type = "Task" [ 770.394618] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.407519] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102656, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.416391] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102652, 'name': CreateVM_Task, 'duration_secs': 1.790685} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.416582] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 770.418180] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.418180] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.418330] env[65107]: WARNING neutronclient.v2_0.client [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.418746] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.418875] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.419988] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 770.420355] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-078b75ea-e574-407a-a3d9-69e8ec48fd05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.425993] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 770.425993] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1372f-59fa-63aa-3633-84d887efa67f" [ 770.425993] env[65107]: _type = "Task" [ 770.425993] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.436549] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1372f-59fa-63aa-3633-84d887efa67f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.552220] env[65107]: WARNING neutronclient.v2_0.client [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.552708] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.553085] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.568917] env[65107]: DEBUG nova.network.neutron [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updated VIF entry in instance network info cache for port 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 770.569283] env[65107]: DEBUG nova.network.neutron [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [{"id": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "address": "fa:16:3e:e0:45:0e", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap257c9dd4-8a", "ovs_interfaceid": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.582785] env[65107]: DEBUG oslo_concurrency.lockutils [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Releasing lock "refresh_cache-bfe544ec-9b08-4118-a940-a51520ecaac0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.583918] env[65107]: DEBUG nova.compute.manager [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Received event network-changed-8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 770.583918] env[65107]: DEBUG nova.compute.manager [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Refreshing instance network info cache due to event network-changed-8622bbd6-6068-43cb-9da2-e996d2202ef2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 770.583918] env[65107]: DEBUG oslo_concurrency.lockutils [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Acquiring lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.583918] env[65107]: DEBUG oslo_concurrency.lockutils [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Acquired lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.583918] env[65107]: DEBUG nova.network.neutron [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Refreshing network info cache for port 8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 770.586194] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ef147d34-69b3-4d17-bc79-ff456c12eba3 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "3948dcdd-e7c1-40b6-a3ec-97741940a0fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.227s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.594891] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102654, 'name': CreateVM_Task, 'duration_secs': 0.588693} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.595157] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 770.595716] env[65107]: WARNING neutronclient.v2_0.client [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.596248] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.721333] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.722396] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.796995] env[65107]: WARNING neutronclient.v2_0.client [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.797698] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.798044] env[65107]: WARNING openstack [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.907082] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102656, 'name': ReconfigVM_Task, 'duration_secs': 0.358921} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.907717] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 19518d90-ba6f-44e5-b7c7-1ececad2b54d/19518d90-ba6f-44e5-b7c7-1ececad2b54d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.908322] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6d32909-c5e0-4f08-8b0f-85be8cab0977 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.916169] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 770.916169] env[65107]: value = "task-5102657" [ 770.916169] env[65107]: _type = "Task" [ 770.916169] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.923231] env[65107]: DEBUG nova.network.neutron [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updated VIF entry in instance network info cache for port 8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 770.923993] env[65107]: DEBUG nova.network.neutron [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [{"id": "776f5e6b-b550-48cb-9141-89cfb94a6460", "address": "fa:16:3e:d2:f3:f6", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.89", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap776f5e6b-b5", "ovs_interfaceid": "776f5e6b-b550-48cb-9141-89cfb94a6460", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5a370d18-6b6f-48cb-b7ea-50951741486f", "address": "fa:16:3e:5d:29:47", "network": {"id": "c771b8d1-00f5-4f8d-93a6-5635a74f7798", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1610194998", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.114", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a370d18-6b", "ovs_interfaceid": "5a370d18-6b6f-48cb-b7ea-50951741486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "address": "fa:16:3e:f4:3b:4f", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5dd2fd-4c", "ovs_interfaceid": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.932153] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102657, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.945193] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1372f-59fa-63aa-3633-84d887efa67f, 'name': SearchDatastore_Task, 'duration_secs': 0.013542} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.946643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.946643] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.947034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.947034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.947276] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.947476] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.947833] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 770.948444] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f35d54f2-1c99-4750-b93f-09e0ff86df2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.951156] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4693eeb6-5380-4cd6-81ae-43bd45b8b7bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.962153] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 770.962153] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241b792-c946-c77e-ff27-93cc4ba2b219" [ 770.962153] env[65107]: _type = "Task" [ 770.962153] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.963721] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.963916] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.968888] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1e5d890-9257-49d4-9543-bee551b94310 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.977725] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 770.977725] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fc3e2-a579-db06-1da0-3d8042664764" [ 770.977725] env[65107]: _type = "Task" [ 770.977725] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.985467] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241b792-c946-c77e-ff27-93cc4ba2b219, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.996316] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522fc3e2-a579-db06-1da0-3d8042664764, 'name': SearchDatastore_Task, 'duration_secs': 0.01279} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.997547] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11d12176-84f2-4b66-816d-fa6958961af4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.010301] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 771.010301] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528afd8d-8d83-e45c-c3c2-5ad8ce943b22" [ 771.010301] env[65107]: _type = "Task" [ 771.010301] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.030172] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528afd8d-8d83-e45c-c3c2-5ad8ce943b22, 'name': SearchDatastore_Task, 'duration_secs': 0.01291} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.031062] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.031494] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 6563cbda-5fd3-4640-9e9a-95904f3f8335/6563cbda-5fd3-4640-9e9a-95904f3f8335.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.031876] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-710de7f3-7ef9-40b3-95bc-8715001552e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.045298] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 771.045298] env[65107]: value = "task-5102658" [ 771.045298] env[65107]: _type = "Task" [ 771.045298] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.056386] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.091986] env[65107]: DEBUG oslo_concurrency.lockutils [req-de9d1700-e315-4339-8df1-7c4f745032af req-f14e9c6e-2da9-4102-ab2b-734a8fdedfb7 service nova] Releasing lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.091986] env[65107]: WARNING neutronclient.v2_0.client [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.091986] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.091986] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.177842] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f957a71b-a580-4c55-aefb-e247b87ea208 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.186506] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ca211f-b612-4882-92e4-9f43b825adb9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.228523] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51366c6-8fb7-450d-873e-1752e9f19e02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.242207] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6ef6d7-7f83-4934-ae7f-fb88b6c5bb72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.262471] env[65107]: DEBUG nova.compute.provider_tree [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 771.280701] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.281075] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.380771] env[65107]: WARNING neutronclient.v2_0.client [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.382643] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.382853] env[65107]: WARNING openstack [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.430495] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102657, 'name': Rename_Task, 'duration_secs': 0.169879} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.430993] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.431258] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b13d2ae1-6f6f-4a0e-9bb8-a59f763ea918 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.433726] env[65107]: DEBUG oslo_concurrency.lockutils [req-641da40f-036f-4ed0-9f09-b18294386cae req-e7749f88-ca13-47c5-9ce0-0c97db47c84c service nova] Releasing lock "refresh_cache-6563cbda-5fd3-4640-9e9a-95904f3f8335" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.447548] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 771.447548] env[65107]: value = "task-5102659" [ 771.447548] env[65107]: _type = "Task" [ 771.447548] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.460355] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102659, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.475535] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241b792-c946-c77e-ff27-93cc4ba2b219, 'name': SearchDatastore_Task, 'duration_secs': 0.021754} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.475940] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.476309] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 771.476537] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.476688] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.476927] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 771.477345] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3e9adba-c232-43fe-b6da-1ab84fb81d02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.491469] env[65107]: DEBUG nova.network.neutron [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updated VIF entry in instance network info cache for port 8622bbd6-6068-43cb-9da2-e996d2202ef2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 771.491869] env[65107]: DEBUG nova.network.neutron [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updating instance_info_cache with network_info: [{"id": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "address": "fa:16:3e:0b:13:8f", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8622bbd6-60", "ovs_interfaceid": "8622bbd6-6068-43cb-9da2-e996d2202ef2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.495018] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 771.495208] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 771.496035] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0676bbcc-cf91-455a-99ed-2a148ddcb880 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.506742] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 771.506742] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e1e0ac-5728-90bc-dc36-4d0562d44e8c" [ 771.506742] env[65107]: _type = "Task" [ 771.506742] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.520341] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e1e0ac-5728-90bc-dc36-4d0562d44e8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.558043] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102658, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.786655] env[65107]: ERROR nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] [req-47d8aaae-6335-44c6-ab2c-7cb984c78336] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-47d8aaae-6335-44c6-ab2c-7cb984c78336"}]} [ 771.810303] env[65107]: DEBUG nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 771.831144] env[65107]: DEBUG nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 771.831494] env[65107]: DEBUG nova.compute.provider_tree [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 771.856386] env[65107]: DEBUG nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 771.887071] env[65107]: DEBUG nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 771.958928] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102659, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.995346] env[65107]: DEBUG oslo_concurrency.lockutils [req-730ec02d-9dac-4d16-b504-986beadffdd8 req-6ecf7ccd-0f8a-4a61-a45a-892d75f2605d service nova] Releasing lock "refresh_cache-82e8c093-b718-4d38-9682-ba2e710e4b93" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.023670] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e1e0ac-5728-90bc-dc36-4d0562d44e8c, 'name': SearchDatastore_Task, 'duration_secs': 0.033982} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.024108] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05e8f567-181d-4677-95ed-f45179c44db0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.031538] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 772.031538] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522744cd-3158-c441-4dfb-5c6d088b92d8" [ 772.031538] env[65107]: _type = "Task" [ 772.031538] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.041396] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522744cd-3158-c441-4dfb-5c6d088b92d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.057625] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533702} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.057982] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 6563cbda-5fd3-4640-9e9a-95904f3f8335/6563cbda-5fd3-4640-9e9a-95904f3f8335.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 772.058147] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.058499] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-36e80912-59db-470c-8578-05b610648d66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.067962] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 772.067962] env[65107]: value = "task-5102660" [ 772.067962] env[65107]: _type = "Task" [ 772.067962] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.080552] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.459808] env[65107]: DEBUG oslo_vmware.api [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102659, 'name': PowerOnVM_Task, 'duration_secs': 0.813235} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.460087] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.460087] env[65107]: INFO nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Took 10.80 seconds to spawn the instance on the hypervisor. [ 772.460253] env[65107]: DEBUG nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 772.461071] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ee025a-e646-4154-8350-cec28907bb29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.488563] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61470d93-6de6-415d-8bcb-8a86b74673e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.498259] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029a81ee-8134-4fae-873f-8695d3d75b33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.538971] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24a5248-8404-464a-ab75-5abca49bfad0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.550265] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522744cd-3158-c441-4dfb-5c6d088b92d8, 'name': SearchDatastore_Task, 'duration_secs': 0.012861} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.552446] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.552713] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c/5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 772.553044] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-823e7a7b-4623-4c02-b5bc-e2b3821687ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.556072] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e380fc8-1189-4183-aa54-37747e8b412d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.572795] env[65107]: DEBUG nova.compute.provider_tree [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 772.578408] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 772.578408] env[65107]: value = "task-5102661" [ 772.578408] env[65107]: _type = "Task" [ 772.578408] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.585506] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.29959} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.586254] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.587118] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a7d295-16f3-48b5-8483-b5b90c3886ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.594013] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.612857] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.613130] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.613339] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.613524] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.613752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.626880] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 6563cbda-5fd3-4640-9e9a-95904f3f8335/6563cbda-5fd3-4640-9e9a-95904f3f8335.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.626880] env[65107]: INFO nova.compute.manager [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Terminating instance [ 772.628768] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9791987-fb54-4c30-8287-f056047e661b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.653659] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 772.653659] env[65107]: value = "task-5102662" [ 772.653659] env[65107]: _type = "Task" [ 772.653659] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.664446] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.708987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "90706f49-af87-4038-ae76-02a95e4a12e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.709280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "90706f49-af87-4038-ae76-02a95e4a12e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.709631] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "90706f49-af87-4038-ae76-02a95e4a12e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.709722] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "90706f49-af87-4038-ae76-02a95e4a12e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.709892] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "90706f49-af87-4038-ae76-02a95e4a12e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.712271] env[65107]: INFO nova.compute.manager [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Terminating instance [ 772.983037] env[65107]: INFO nova.compute.manager [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Took 42.82 seconds to build instance. [ 773.104645] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102661, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.129173] env[65107]: DEBUG nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 773.129521] env[65107]: DEBUG nova.compute.provider_tree [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 69 to 70 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 773.129815] env[65107]: DEBUG nova.compute.provider_tree [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 773.146623] env[65107]: DEBUG nova.compute.manager [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 773.146722] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.147679] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67117c40-87ab-4e75-ba4d-055edbb77209 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.165963] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 773.167058] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87c012b0-ec38-44e8-8671-da85c3bdcda3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.177067] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102662, 'name': ReconfigVM_Task, 'duration_secs': 0.425162} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.178020] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 6563cbda-5fd3-4640-9e9a-95904f3f8335/6563cbda-5fd3-4640-9e9a-95904f3f8335.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.179059] env[65107]: DEBUG oslo_vmware.api [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 773.179059] env[65107]: value = "task-5102663" [ 773.179059] env[65107]: _type = "Task" [ 773.179059] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.179343] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-54138c55-9a29-4fc3-bf3b-9cdca5b0ffaa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.193390] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 773.193390] env[65107]: value = "task-5102664" [ 773.193390] env[65107]: _type = "Task" [ 773.193390] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.197261] env[65107]: DEBUG oslo_vmware.api [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.209501] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102664, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.217999] env[65107]: DEBUG nova.compute.manager [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 773.217999] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.218584] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fcdc57-86e0-407a-b443-087e1feb8ba6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.231987] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 773.234650] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb3cbac0-dbbe-4cdb-ac38-01c285f91438 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.243458] env[65107]: DEBUG oslo_vmware.api [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 773.243458] env[65107]: value = "task-5102665" [ 773.243458] env[65107]: _type = "Task" [ 773.243458] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.257812] env[65107]: DEBUG oslo_vmware.api [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.376779] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.485217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b77ca92-030d-47f6-b62c-e22daee396ec tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.604s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.486150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.110s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.486718] env[65107]: DEBUG nova.compute.manager [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 773.487877] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76719821-0c8a-4a7d-bdde-5b6b40311375 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.496671] env[65107]: DEBUG nova.compute.manager [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 773.497472] env[65107]: DEBUG nova.objects.instance [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lazy-loading 'flavor' on Instance uuid 19518d90-ba6f-44e5-b7c7-1ececad2b54d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.593672] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102661, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.638962] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.625s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.641805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.910s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.643515] env[65107]: INFO nova.compute.claims [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.667718] env[65107]: INFO nova.scheduler.client.report [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Deleted allocations for instance b20a2c71-514d-488a-8edb-8fb0274a894a [ 773.699035] env[65107]: DEBUG oslo_vmware.api [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102663, 'name': PowerOffVM_Task, 'duration_secs': 0.257056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.699035] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.699035] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.699035] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1a5304b-2e4e-4e79-a323-9352601395a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.710849] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102664, 'name': Rename_Task, 'duration_secs': 0.270438} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.710982] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.711277] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89557edc-c0e4-4605-a12c-73584a4283ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.719797] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 773.719797] env[65107]: value = "task-5102667" [ 773.719797] env[65107]: _type = "Task" [ 773.719797] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.729807] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.756728] env[65107]: DEBUG oslo_vmware.api [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102665, 'name': PowerOffVM_Task, 'duration_secs': 0.236214} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.757225] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.757418] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.757728] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d422dd70-8c76-4334-bcdb-6c573d41bad1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.776685] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.776936] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.777238] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleting the datastore file [datastore1] c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.777587] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f9367fb-ee33-4d2d-b53a-1adff312a7ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.784933] env[65107]: DEBUG oslo_vmware.api [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 773.784933] env[65107]: value = "task-5102669" [ 773.784933] env[65107]: _type = "Task" [ 773.784933] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.797101] env[65107]: DEBUG oslo_vmware.api [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.841332] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.841604] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.841823] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleting the datastore file [datastore1] 90706f49-af87-4038-ae76-02a95e4a12e6 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.842162] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-894da251-bc6a-48e0-b8d4-3bb6508cfb70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.855404] env[65107]: DEBUG oslo_vmware.api [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for the task: (returnval){ [ 773.855404] env[65107]: value = "task-5102670" [ 773.855404] env[65107]: _type = "Task" [ 773.855404] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.866912] env[65107]: DEBUG oslo_vmware.api [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.095368] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102661, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.179719] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2d1b03b-da5c-4c7b-abc1-124064a3cd1b tempest-ServerRescueTestJSONUnderV235-1815693667 tempest-ServerRescueTestJSONUnderV235-1815693667-project-member] Lock "b20a2c71-514d-488a-8edb-8fb0274a894a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.010s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.231505] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102667, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.296740] env[65107]: DEBUG oslo_vmware.api [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200598} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.296892] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.297092] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 774.297284] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.297473] env[65107]: INFO nova.compute.manager [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Took 1.15 seconds to destroy the instance on the hypervisor. [ 774.297753] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 774.298111] env[65107]: DEBUG nova.compute.manager [-] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 774.298111] env[65107]: DEBUG nova.network.neutron [-] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 774.298864] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.299195] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.299195] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.367825] env[65107]: DEBUG oslo_vmware.api [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Task: {'id': task-5102670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191414} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.369165] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.371321] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.371621] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 774.371938] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.372110] env[65107]: INFO nova.compute.manager [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 774.372414] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 774.373131] env[65107]: DEBUG nova.compute.manager [-] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 774.374094] env[65107]: DEBUG nova.network.neutron [-] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 774.374094] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.374553] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.374674] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.512019] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 774.512019] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a14a3276-580b-4277-9655-daf46656c9ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.517191] env[65107]: DEBUG oslo_vmware.api [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 774.517191] env[65107]: value = "task-5102671" [ 774.517191] env[65107]: _type = "Task" [ 774.517191] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.527144] env[65107]: DEBUG oslo_vmware.api [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.596938] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102661, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.905617} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.598106] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c/5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.598317] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.598618] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f75d4c0-5be7-4ff0-8652-608e1b0de32a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.608517] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 774.608517] env[65107]: value = "task-5102672" [ 774.608517] env[65107]: _type = "Task" [ 774.608517] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.619829] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.659021] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.735135] env[65107]: DEBUG oslo_vmware.api [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102667, 'name': PowerOnVM_Task, 'duration_secs': 0.576004} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.735489] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.735793] env[65107]: INFO nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Took 16.01 seconds to spawn the instance on the hypervisor. [ 774.737992] env[65107]: DEBUG nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 774.737992] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8041a9-2c63-4d46-b53f-6488c189315c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.032258] env[65107]: DEBUG oslo_vmware.api [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102671, 'name': PowerOffVM_Task, 'duration_secs': 0.344554} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.032541] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.032736] env[65107]: DEBUG nova.compute.manager [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 775.033598] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c45bb0-5d56-4a19-a7d9-5cf31d91e770 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.122352] env[65107]: DEBUG nova.network.neutron [-] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 775.136679] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11141} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.138474] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.142321] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390d1967-750f-465f-838f-4b4bcc6f6816 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.172755] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c/5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.176076] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03d4e36d-21a4-4702-8d10-307f9f84b095 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.198444] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 775.198444] env[65107]: value = "task-5102673" [ 775.198444] env[65107]: _type = "Task" [ 775.198444] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.214194] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102673, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.238499] env[65107]: DEBUG nova.compute.manager [req-99df9890-e480-42f3-9246-a9e7613e1b03 req-c6b5f1f7-798a-40d8-b288-2ff823dfae46 service nova] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Received event network-vif-deleted-7b78f034-800f-47c1-9c6f-7e94df8703b1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 775.262231] env[65107]: INFO nova.compute.manager [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Took 46.28 seconds to build instance. [ 775.365979] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5a2bcf-c755-4540-af1c-e982cb168438 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.374606] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288b7eb4-4e16-4e77-bfa1-ed9539311f2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.417660] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c4eded-85d0-462c-a0f5-7405ba62d340 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.428335] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5bc3a0-bcbe-480b-ac23-546766ff37d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.445137] env[65107]: DEBUG nova.network.neutron [-] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 775.448017] env[65107]: DEBUG nova.compute.provider_tree [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.533881] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "5fccc802-f5ac-4450-8246-4cf9a5371046" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.535673] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.547420] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebdcc64d-f125-4c3b-9720-fe6dcfc1b075 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.060s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.639171] env[65107]: INFO nova.compute.manager [-] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Took 1.34 seconds to deallocate network for instance. [ 775.713196] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102673, 'name': ReconfigVM_Task, 'duration_secs': 0.301958} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.713196] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c/5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.713196] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c13cbd7-7825-40e8-86bd-d80a561e758a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.721903] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 775.721903] env[65107]: value = "task-5102674" [ 775.721903] env[65107]: _type = "Task" [ 775.721903] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.730792] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102674, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.766026] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ebbc4e7c-74b8-47df-bb87-0f5dcfe90c91 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.315s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.950458] env[65107]: INFO nova.compute.manager [-] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Took 1.58 seconds to deallocate network for instance. [ 775.954915] env[65107]: DEBUG nova.scheduler.client.report [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 776.039083] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 776.146085] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.234499] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102674, 'name': Rename_Task, 'duration_secs': 0.183133} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.234814] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.235076] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6de32a3-6a75-4eb1-bd4a-6ae47ad6b91d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.243544] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 776.243544] env[65107]: value = "task-5102675" [ 776.243544] env[65107]: _type = "Task" [ 776.243544] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.252968] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.450217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "6563cbda-5fd3-4640-9e9a-95904f3f8335" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.450217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.450345] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.450534] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.450708] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.454049] env[65107]: INFO nova.compute.manager [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Terminating instance [ 776.464025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.464377] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.823s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.464981] env[65107]: DEBUG nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 776.469396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.140s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.469652] env[65107]: DEBUG nova.objects.instance [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lazy-loading 'resources' on Instance uuid 854f7a3f-9b7a-4506-8310-94533f0e4cfc {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 776.567897] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.755918] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102675, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.958435] env[65107]: DEBUG nova.compute.manager [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 776.958867] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.960048] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b223fb7b-a630-4be0-bc7a-df6447f119f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.968563] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 776.969686] env[65107]: DEBUG nova.compute.utils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 776.972257] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-540302ae-5c53-4956-8b05-520a272e93a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.974988] env[65107]: DEBUG nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 776.987386] env[65107]: DEBUG oslo_vmware.api [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 776.987386] env[65107]: value = "task-5102676" [ 776.987386] env[65107]: _type = "Task" [ 776.987386] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.999721] env[65107]: DEBUG oslo_vmware.api [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.260810] env[65107]: DEBUG oslo_vmware.api [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102675, 'name': PowerOnVM_Task, 'duration_secs': 0.526103} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.260961] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.261207] env[65107]: INFO nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Took 10.69 seconds to spawn the instance on the hypervisor. [ 777.261486] env[65107]: DEBUG nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 777.262524] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811173da-0499-4d03-a90d-60923dab3634 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.281572] env[65107]: DEBUG nova.compute.manager [req-f2083c99-e0a1-440a-80d5-975651d6ce61 req-50a2910d-c468-4501-a149-5566afceb192 service nova] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Received event network-vif-deleted-52873450-f314-4110-a316-7380569e0afe {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 777.327721] env[65107]: DEBUG nova.compute.manager [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 777.327921] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049b1e4f-3d04-4383-9a28-3a52bb6def41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.473541] env[65107]: DEBUG nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 777.498861] env[65107]: DEBUG oslo_vmware.api [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102676, 'name': PowerOffVM_Task, 'duration_secs': 0.214761} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.498861] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 777.498861] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 777.499121] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0e4208a-1ad4-4bcc-9d19-fae603ac6fc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.505316] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d22792e-7bca-4332-9275-23fe50170579 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.513778] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa51a06-8cdf-4af7-9c82-f775114ef480 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.545090] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3804ad9b-27bc-4c33-91cf-e8eb753bc896 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.553175] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285d787a-6068-441f-8474-cbd4fc5548d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.568981] env[65107]: DEBUG nova.compute.provider_tree [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.705933] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 777.705933] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 777.705933] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Deleting the datastore file [datastore2] 6563cbda-5fd3-4640-9e9a-95904f3f8335 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.706171] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fced3e00-dfba-4385-ad7c-9c56dc189b14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.713177] env[65107]: DEBUG oslo_vmware.api [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 777.713177] env[65107]: value = "task-5102678" [ 777.713177] env[65107]: _type = "Task" [ 777.713177] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.721479] env[65107]: DEBUG oslo_vmware.api [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.784181] env[65107]: INFO nova.compute.manager [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Took 44.32 seconds to build instance. [ 777.843652] env[65107]: INFO nova.compute.manager [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] instance snapshotting [ 777.844268] env[65107]: WARNING nova.compute.manager [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 777.847138] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94d23c9-2a04-41b0-aee9-0310d00bb337 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.877125] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad291b57-ddd6-4cc9-87e6-3fb67b686426 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.076027] env[65107]: DEBUG nova.scheduler.client.report [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 778.225186] env[65107]: DEBUG oslo_vmware.api [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.368416} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.226691] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 778.226691] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 778.226691] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.226691] env[65107]: INFO nova.compute.manager [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Took 1.27 seconds to destroy the instance on the hypervisor. [ 778.226691] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 778.226691] env[65107]: DEBUG nova.compute.manager [-] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 778.226906] env[65107]: DEBUG nova.network.neutron [-] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 778.226992] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.227558] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.227791] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.289090] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.292472] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e1b7a48-5c0c-4934-929d-fc410c3a715e tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.520s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.389310] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 778.389646] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-92f62f3f-16a7-4900-8b1b-99ae725f38d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.397849] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 778.397849] env[65107]: value = "task-5102679" [ 778.397849] env[65107]: _type = "Task" [ 778.397849] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.410278] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102679, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.484106] env[65107]: DEBUG nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 778.519191] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 778.519533] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 778.519699] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 778.519882] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 778.520131] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 778.520182] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 778.520390] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.520578] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 778.520723] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 778.520885] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 778.521059] env[65107]: DEBUG nova.virt.hardware [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 778.522461] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c171a583-15a4-4ac5-8816-58eed1169a28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.533024] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bd960e-b3ac-4d57-8aff-3f5a91a4e9ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.552793] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.560225] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Creating folder: Project (e01d600d7e6b4e39bc6c2f154f8781f0). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.560914] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebe08179-abdf-443d-ae6d-25023ddeb473 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.577405] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Created folder: Project (e01d600d7e6b4e39bc6c2f154f8781f0) in parent group-v992574. [ 778.577756] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Creating folder: Instances. Parent ref: group-v992709. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.578214] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b51b0dc-d176-4053-beed-7a4de494d1bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.580913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.111s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.584650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.562s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.585980] env[65107]: DEBUG nova.objects.instance [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lazy-loading 'resources' on Instance uuid ad53146e-210d-4321-b819-90795c1af260 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.601410] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Created folder: Instances in parent group-v992709. [ 778.601678] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 778.601887] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.602319] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94b9e5f2-ced4-43dc-9dd9-f54a0c3e01f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.622575] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.622575] env[65107]: value = "task-5102682" [ 778.622575] env[65107]: _type = "Task" [ 778.622575] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.627944] env[65107]: INFO nova.scheduler.client.report [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Deleted allocations for instance 854f7a3f-9b7a-4506-8310-94533f0e4cfc [ 778.639719] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102682, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.913272] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102679, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.040238] env[65107]: DEBUG nova.compute.manager [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 779.040478] env[65107]: DEBUG nova.compute.manager [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing instance network info cache due to event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 779.040729] env[65107]: DEBUG oslo_concurrency.lockutils [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Acquiring lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.040925] env[65107]: DEBUG oslo_concurrency.lockutils [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Acquired lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.041282] env[65107]: DEBUG nova.network.neutron [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 779.137025] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102682, 'name': CreateVM_Task, 'duration_secs': 0.305373} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.138498] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.138934] env[65107]: DEBUG oslo_concurrency.lockutils [None req-34c104bc-34d1-4a35-afdd-53cb78b8a6e0 tempest-ServerDiagnosticsTest-914305661 tempest-ServerDiagnosticsTest-914305661-project-member] Lock "854f7a3f-9b7a-4506-8310-94533f0e4cfc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.310s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.143048] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.143048] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.143048] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 779.143048] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03cb5d05-c350-4586-a21e-1d38cf676a9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.151032] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 779.151032] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529dd973-dd58-56f4-80b4-f68a12ef7364" [ 779.151032] env[65107]: _type = "Task" [ 779.151032] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.162806] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529dd973-dd58-56f4-80b4-f68a12ef7364, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.409801] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102679, 'name': CreateSnapshot_Task, 'duration_secs': 0.548913} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.412850] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 779.414124] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67087f3f-def8-41a5-a087-2459ffeddc39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.454796] env[65107]: DEBUG nova.network.neutron [-] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 779.470302] env[65107]: DEBUG nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-vif-deleted-776f5e6b-b550-48cb-9141-89cfb94a6460 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 779.470302] env[65107]: INFO nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Neutron deleted interface 776f5e6b-b550-48cb-9141-89cfb94a6460; detaching it from the instance and deleting it from the info cache [ 779.470302] env[65107]: DEBUG nova.network.neutron [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [{"id": "5a370d18-6b6f-48cb-b7ea-50951741486f", "address": "fa:16:3e:5d:29:47", "network": {"id": "c771b8d1-00f5-4f8d-93a6-5635a74f7798", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1610194998", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.114", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a370d18-6b", "ovs_interfaceid": "5a370d18-6b6f-48cb-b7ea-50951741486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "address": "fa:16:3e:f4:3b:4f", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5dd2fd-4c", "ovs_interfaceid": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 779.545132] env[65107]: WARNING neutronclient.v2_0.client [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 779.545958] env[65107]: WARNING openstack [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.546340] env[65107]: WARNING openstack [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.625805] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c2f034-e299-42c3-8d62-27aca7a1848b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.635623] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d858abe-d7e6-488f-9520-0a27aa7b8f99 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.679133] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3a0ca5-8061-48cb-814b-823a6eede0ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.690945] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529dd973-dd58-56f4-80b4-f68a12ef7364, 'name': SearchDatastore_Task, 'duration_secs': 0.033572} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.692214] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2ef2d0-3be2-4779-ab6e-86ecb1e96690 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.699601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.699601] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.699601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.699601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.699601] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.699601] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06b90eea-a239-43c6-8a8d-d63e09990adf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.711363] env[65107]: DEBUG nova.compute.provider_tree [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.713847] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.714024] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.715018] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a97b5714-b356-4fe3-a642-940c3e7db665 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.722178] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 779.722178] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5284f8c0-9100-1325-80c2-0919eeeeeecd" [ 779.722178] env[65107]: _type = "Task" [ 779.722178] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.728317] env[65107]: WARNING openstack [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.728798] env[65107]: WARNING openstack [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.742385] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5284f8c0-9100-1325-80c2-0919eeeeeecd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.927187] env[65107]: WARNING neutronclient.v2_0.client [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 779.927922] env[65107]: WARNING openstack [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.928364] env[65107]: WARNING openstack [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.947360] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 779.948044] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b690e023-c8db-4628-a51d-32c6f5ea8387 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.957721] env[65107]: INFO nova.compute.manager [-] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Took 1.73 seconds to deallocate network for instance. [ 779.958293] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 779.958293] env[65107]: value = "task-5102683" [ 779.958293] env[65107]: _type = "Task" [ 779.958293] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.973287] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2471d53c-9a0c-4cd1-94c7-8c6f20e4e631 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.978826] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102683, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.989718] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b613806-0f9a-40cc-afc4-6efb0de3b4b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.037571] env[65107]: DEBUG nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Detach interface failed, port_id=776f5e6b-b550-48cb-9141-89cfb94a6460, reason: Instance 6563cbda-5fd3-4640-9e9a-95904f3f8335 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 780.037571] env[65107]: DEBUG nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-vif-deleted-5a370d18-6b6f-48cb-b7ea-50951741486f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 780.037571] env[65107]: INFO nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Neutron deleted interface 5a370d18-6b6f-48cb-b7ea-50951741486f; detaching it from the instance and deleting it from the info cache [ 780.037571] env[65107]: DEBUG nova.network.neutron [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [{"id": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "address": "fa:16:3e:f4:3b:4f", "network": {"id": "0aaf8b58-1398-4ccf-b1e9-f4f02cbec0f9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1573198494", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.247", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1b8b991-feba-44e6-900c-6486e7e122f0", "external-id": "nsx-vlan-transportzone-429", "segmentation_id": 429, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f5dd2fd-4c", "ovs_interfaceid": "8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.183096] env[65107]: DEBUG nova.network.neutron [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updated VIF entry in instance network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 780.183608] env[65107]: DEBUG nova.network.neutron [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.216554] env[65107]: DEBUG nova.scheduler.client.report [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.234548] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5284f8c0-9100-1325-80c2-0919eeeeeecd, 'name': SearchDatastore_Task, 'duration_secs': 0.030358} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.236033] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17e3e534-5f79-434f-83ba-7cd8d9d923eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.247272] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 780.247272] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529781c2-e75e-a878-2f4c-9366af6efa74" [ 780.247272] env[65107]: _type = "Task" [ 780.247272] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.257954] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529781c2-e75e-a878-2f4c-9366af6efa74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.473592] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.477481] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102683, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.544593] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ff1ed98-69b4-4cf8-b9c6-3fa78af1958e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.563157] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0ce90a-d55f-44e9-9d7a-870ea050ddbb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.605788] env[65107]: DEBUG nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Detach interface failed, port_id=5a370d18-6b6f-48cb-b7ea-50951741486f, reason: Instance 6563cbda-5fd3-4640-9e9a-95904f3f8335 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 780.606693] env[65107]: DEBUG nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Received event network-vif-deleted-8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 780.606796] env[65107]: INFO nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Neutron deleted interface 8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138; detaching it from the instance and deleting it from the info cache [ 780.607027] env[65107]: DEBUG nova.network.neutron [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.686751] env[65107]: DEBUG oslo_concurrency.lockutils [req-f920df41-599b-4410-b673-ae2c85e77c65 req-1940001f-e4a4-4a00-b8bf-bc982747841b service nova] Releasing lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.721576] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.138s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.724216] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.590s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.725991] env[65107]: INFO nova.compute.claims [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.761638] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529781c2-e75e-a878-2f4c-9366af6efa74, 'name': SearchDatastore_Task, 'duration_secs': 0.023788} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.761638] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.761925] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.762503] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e1e70e4-7af3-42ce-9911-84419487881c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.766323] env[65107]: INFO nova.scheduler.client.report [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted allocations for instance ad53146e-210d-4321-b819-90795c1af260 [ 780.778798] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 780.778798] env[65107]: value = "task-5102684" [ 780.778798] env[65107]: _type = "Task" [ 780.778798] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.781817] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.979427] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102683, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.110982] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b68705a6-713f-44b9-ac1d-1b201d746830 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.127465] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6501c1-d863-4553-9420-bbaa34af71cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.175875] env[65107]: DEBUG nova.compute.manager [req-0f809b27-75a1-4180-95b9-83092c7ac965 req-d073305b-40f3-48c3-b554-ad35549e1d8a service nova] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Detach interface failed, port_id=8f5dd2fd-4cf5-44d2-ad85-4d9916e1b138, reason: Instance 6563cbda-5fd3-4640-9e9a-95904f3f8335 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 781.280564] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e2406e37-4ab6-46db-9261-f22973f29dc3 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "ad53146e-210d-4321-b819-90795c1af260" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.841s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.290521] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102684, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.478524] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102683, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.791019] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676493} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.796238] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.796543] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.797340] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-889103df-412e-48c5-8665-b26584b13f8d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.804660] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 781.804660] env[65107]: value = "task-5102685" [ 781.804660] env[65107]: _type = "Task" [ 781.804660] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.813948] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102685, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.969477] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "f58c3088-c821-4eec-be0a-81221debc98e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 781.969675] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "f58c3088-c821-4eec-be0a-81221debc98e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.970017] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "f58c3088-c821-4eec-be0a-81221debc98e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 781.971421] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "f58c3088-c821-4eec-be0a-81221debc98e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.972712] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "f58c3088-c821-4eec-be0a-81221debc98e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.981198] env[65107]: INFO nova.compute.manager [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Terminating instance [ 781.994218] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102683, 'name': CloneVM_Task, 'duration_secs': 2.018017} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.995197] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Created linked-clone VM from snapshot [ 781.996533] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55078204-7893-44c0-8048-000b76d88e9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.009593] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Uploading image cbb434d5-e178-436d-a497-b20fc87120c6 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 782.041828] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 782.041828] env[65107]: value = "vm-992713" [ 782.041828] env[65107]: _type = "VirtualMachine" [ 782.041828] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 782.042956] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-01c61060-494c-4a37-bbe7-93d745139d62 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.050954] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease: (returnval){ [ 782.050954] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529312aa-4e72-306d-e176-9bf761c8c573" [ 782.050954] env[65107]: _type = "HttpNfcLease" [ 782.050954] env[65107]: } obtained for exporting VM: (result){ [ 782.050954] env[65107]: value = "vm-992713" [ 782.050954] env[65107]: _type = "VirtualMachine" [ 782.050954] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 782.051179] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the lease: (returnval){ [ 782.051179] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529312aa-4e72-306d-e176-9bf761c8c573" [ 782.051179] env[65107]: _type = "HttpNfcLease" [ 782.051179] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 782.059931] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 782.059931] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529312aa-4e72-306d-e176-9bf761c8c573" [ 782.059931] env[65107]: _type = "HttpNfcLease" [ 782.059931] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 782.094344] env[65107]: DEBUG nova.compute.manager [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 782.094560] env[65107]: DEBUG nova.compute.manager [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing instance network info cache due to event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 782.094845] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Acquiring lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.094997] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Acquired lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.095172] env[65107]: DEBUG nova.network.neutron [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 782.317798] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102685, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078171} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.318486] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.320396] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e80632-5235-4290-ad16-3a24ad03676d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.348081] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.351118] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f572c0a-abeb-4b9e-9ae9-0c64c3938879 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.372929] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 782.372929] env[65107]: value = "task-5102687" [ 782.372929] env[65107]: _type = "Task" [ 782.372929] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.387380] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102687, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.399107] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c69187-e090-44f9-b881-acbc8f659caf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.408901] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8e7179-87fe-4054-996f-62aab627d137 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.444821] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2fe4b7-d0e8-40a3-91ca-157ac6b9d233 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.449025] env[65107]: DEBUG nova.compute.manager [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Received event network-changed-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 782.449025] env[65107]: DEBUG nova.compute.manager [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Refreshing instance network info cache due to event network-changed-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 782.449368] env[65107]: DEBUG oslo_concurrency.lockutils [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Acquiring lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.449583] env[65107]: DEBUG oslo_concurrency.lockutils [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Acquired lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.449805] env[65107]: DEBUG nova.network.neutron [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Refreshing network info cache for port 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 782.457667] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf04d611-7cbd-4cf0-86ca-547febdae846 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.478340] env[65107]: DEBUG nova.compute.provider_tree [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.487014] env[65107]: DEBUG nova.compute.manager [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 782.487234] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.488618] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842fd012-66c9-4f32-85fe-bccb00410bb2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.498515] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.498854] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c234e09-3942-4c21-85dc-16a8397e890a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.508029] env[65107]: DEBUG oslo_vmware.api [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 782.508029] env[65107]: value = "task-5102688" [ 782.508029] env[65107]: _type = "Task" [ 782.508029] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.519237] env[65107]: DEBUG oslo_vmware.api [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.564501] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 782.564501] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529312aa-4e72-306d-e176-9bf761c8c573" [ 782.564501] env[65107]: _type = "HttpNfcLease" [ 782.564501] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 782.566092] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 782.566092] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529312aa-4e72-306d-e176-9bf761c8c573" [ 782.566092] env[65107]: _type = "HttpNfcLease" [ 782.566092] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 782.566092] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bd6505-ade0-402e-ae91-034c3524bef0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.575804] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522ad92e-31de-403d-5636-0663b99a8799/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 782.575961] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522ad92e-31de-403d-5636-0663b99a8799/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 782.641278] env[65107]: WARNING neutronclient.v2_0.client [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.642191] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.642842] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.763065] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-48519676-570a-41ec-8eb4-b2d21a084974 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.777502] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.777929] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.886205] env[65107]: WARNING neutronclient.v2_0.client [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.887105] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.887297] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.898920] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.954253] env[65107]: WARNING neutronclient.v2_0.client [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.955651] env[65107]: WARNING openstack [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.956629] env[65107]: WARNING openstack [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.981958] env[65107]: DEBUG nova.scheduler.client.report [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 783.020889] env[65107]: DEBUG oslo_vmware.api [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102688, 'name': PowerOffVM_Task, 'duration_secs': 0.377188} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.021325] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.021418] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 783.021619] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e830f6e-5370-45d0-a0b8-956e98945d8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.026880] env[65107]: DEBUG nova.network.neutron [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updated VIF entry in instance network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 783.027265] env[65107]: DEBUG nova.network.neutron [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.094361] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 783.094568] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 783.095027] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleting the datastore file [datastore2] f58c3088-c821-4eec-be0a-81221debc98e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 783.095154] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc5334ae-83e1-47b1-98e5-83578c17304f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.106025] env[65107]: DEBUG oslo_vmware.api [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 783.106025] env[65107]: value = "task-5102690" [ 783.106025] env[65107]: _type = "Task" [ 783.106025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.117464] env[65107]: DEBUG oslo_vmware.api [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.201691] env[65107]: WARNING openstack [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.202772] env[65107]: WARNING openstack [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.305683] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.307120] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.308604] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.309224] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.309224] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.314644] env[65107]: WARNING neutronclient.v2_0.client [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.315370] env[65107]: WARNING openstack [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.315808] env[65107]: WARNING openstack [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.326808] env[65107]: INFO nova.compute.manager [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Terminating instance [ 783.396555] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102687, 'name': ReconfigVM_Task, 'duration_secs': 0.866944} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.396555] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.396555] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71231373-fd05-4632-8693-37864ef0b2ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.405590] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 783.405590] env[65107]: value = "task-5102691" [ 783.405590] env[65107]: _type = "Task" [ 783.405590] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.419289] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102691, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.438355] env[65107]: DEBUG nova.network.neutron [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updated VIF entry in instance network info cache for port 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 783.438685] env[65107]: DEBUG nova.network.neutron [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [{"id": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "address": "fa:16:3e:e0:45:0e", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap257c9dd4-8a", "ovs_interfaceid": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.487794] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.764s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.490743] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 783.494825] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.841s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.496067] env[65107]: DEBUG nova.objects.instance [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lazy-loading 'resources' on Instance uuid be518c1d-edd7-40ff-b7cc-3310885b07b4 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 783.531880] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Releasing lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.531880] env[65107]: DEBUG nova.compute.manager [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Received event network-changed-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 783.531880] env[65107]: DEBUG nova.compute.manager [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Refreshing instance network info cache due to event network-changed-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 783.532186] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Acquiring lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.627382] env[65107]: DEBUG oslo_vmware.api [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321068} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.627978] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 783.628303] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 783.628440] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 783.628726] env[65107]: INFO nova.compute.manager [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 783.629588] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 783.629588] env[65107]: DEBUG nova.compute.manager [-] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 783.629588] env[65107]: DEBUG nova.network.neutron [-] [instance: f58c3088-c821-4eec-be0a-81221debc98e] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 783.630191] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.631139] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.631615] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.838438] env[65107]: DEBUG nova.compute.manager [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 783.838438] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.838438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280d8a1e-3c42-4bf1-8de0-e9b8067f5616 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.849184] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.852513] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.852513] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae53a52f-fd86-4599-8cd1-eb68f3c3ce3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.861457] env[65107]: DEBUG oslo_vmware.api [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 783.861457] env[65107]: value = "task-5102692" [ 783.861457] env[65107]: _type = "Task" [ 783.861457] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.873945] env[65107]: DEBUG oslo_vmware.api [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.916695] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102691, 'name': Rename_Task, 'duration_secs': 0.237255} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.917200] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 783.917200] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94e1ba08-ef41-4dc4-8eec-7737a2fb1980 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.932449] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 783.932449] env[65107]: value = "task-5102693" [ 783.932449] env[65107]: _type = "Task" [ 783.932449] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.944278] env[65107]: DEBUG oslo_concurrency.lockutils [req-c42541c4-6939-4b3b-98b8-b5e851c67823 req-3012f237-b981-415b-9b74-71af169de1bd service nova] Releasing lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.945162] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Acquired lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.945162] env[65107]: DEBUG nova.network.neutron [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Refreshing network info cache for port 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 783.951664] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.996642] env[65107]: DEBUG nova.compute.utils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 783.999020] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 783.999398] env[65107]: DEBUG nova.network.neutron [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 784.003400] env[65107]: WARNING neutronclient.v2_0.client [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 784.003869] env[65107]: WARNING neutronclient.v2_0.client [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 784.004516] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 784.004857] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 784.180722] env[65107]: DEBUG nova.policy [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e7e78dbbfc041da983b34a950e383b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '634f89d91e1443c7a0eb1314ecb17232', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 784.378443] env[65107]: DEBUG oslo_vmware.api [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102692, 'name': PowerOffVM_Task, 'duration_secs': 0.331359} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.379174] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.379508] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 784.379904] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0a09e2e-0a36-4ee7-bc1e-ecacdda8dc42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.447971] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102693, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.453203] env[65107]: WARNING neutronclient.v2_0.client [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 784.454565] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 784.454713] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 784.466320] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 784.466441] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 784.466593] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Deleting the datastore file [datastore2] 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.466889] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-597f089f-95b1-4eb2-a13c-44543831fe07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.476969] env[65107]: DEBUG oslo_vmware.api [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 784.476969] env[65107]: value = "task-5102695" [ 784.476969] env[65107]: _type = "Task" [ 784.476969] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.491840] env[65107]: DEBUG oslo_vmware.api [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.516534] env[65107]: DEBUG nova.compute.utils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 784.637048] env[65107]: DEBUG nova.network.neutron [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Successfully created port: a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 784.645019] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1744af-dfa1-42c5-8208-046b4a4b352e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.653919] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bda5186-6b36-4cc1-aa95-887eaee9c7be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.691367] env[65107]: DEBUG nova.network.neutron [-] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 784.692372] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1892a8c-49e5-41ff-8ccf-b525cca05c1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.701967] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5c0008-920a-4f60-b83a-60e71a1907c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.719023] env[65107]: DEBUG nova.compute.provider_tree [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.947373] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102693, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.989372] env[65107]: DEBUG oslo_vmware.api [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305344} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.989712] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.989901] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.990093] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.990268] env[65107]: INFO nova.compute.manager [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 784.990797] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 784.991026] env[65107]: DEBUG nova.compute.manager [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 784.991130] env[65107]: DEBUG nova.network.neutron [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 784.991393] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 784.992269] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 784.992385] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.019425] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 785.192998] env[65107]: INFO nova.compute.manager [-] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Took 1.56 seconds to deallocate network for instance. [ 785.222426] env[65107]: DEBUG nova.scheduler.client.report [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 785.448787] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102693, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.702112] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.728945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.234s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.731618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.860s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.731901] env[65107]: DEBUG nova.objects.instance [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lazy-loading 'resources' on Instance uuid d40aec5a-d898-42f3-a121-503cf043f72b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.754696] env[65107]: INFO nova.scheduler.client.report [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Deleted allocations for instance be518c1d-edd7-40ff-b7cc-3310885b07b4 [ 785.947185] env[65107]: DEBUG oslo_vmware.api [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102693, 'name': PowerOnVM_Task, 'duration_secs': 1.541237} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.947529] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.947984] env[65107]: INFO nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Took 7.46 seconds to spawn the instance on the hypervisor. [ 785.947984] env[65107]: DEBUG nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 785.948873] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8d8ccb-6718-4a37-a646-9e0e616bb2e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.031165] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 786.064166] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:21:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='379790923',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-1970379414',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 786.064448] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 786.064595] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 786.064768] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 786.064915] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 786.065086] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 786.065302] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.065534] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 786.065900] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 786.065974] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 786.066149] env[65107]: DEBUG nova.virt.hardware [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 786.067187] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cfacf2-1c0f-4bae-8612-1b84de6c50fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.077035] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272df71a-c512-4904-8a6b-55cc5f946b8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.267742] env[65107]: DEBUG oslo_concurrency.lockutils [None req-18520d22-1dba-4263-ad31-f2acde5b01f8 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "be518c1d-edd7-40ff-b7cc-3310885b07b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.471s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.311335] env[65107]: DEBUG nova.network.neutron [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Successfully updated port: a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 786.432712] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.433122] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.442153] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.471721] env[65107]: INFO nova.compute.manager [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Took 46.77 seconds to build instance. [ 786.559868] env[65107]: WARNING neutronclient.v2_0.client [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.560595] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.560948] env[65107]: WARNING openstack [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.675936] env[65107]: DEBUG nova.compute.manager [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 786.676029] env[65107]: DEBUG nova.compute.manager [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing instance network info cache due to event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 786.676240] env[65107]: DEBUG oslo_concurrency.lockutils [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Acquiring lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.676374] env[65107]: DEBUG oslo_concurrency.lockutils [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Acquired lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.676529] env[65107]: DEBUG nova.network.neutron [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 786.690104] env[65107]: DEBUG nova.network.neutron [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updated VIF entry in instance network info cache for port 257c9dd4-8ac7-4bca-abb6-a32d3652c6a2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 786.690987] env[65107]: DEBUG nova.network.neutron [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [{"id": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "address": "fa:16:3e:e0:45:0e", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap257c9dd4-8a", "ovs_interfaceid": "257c9dd4-8ac7-4bca-abb6-a32d3652c6a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 786.821702] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.821789] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.821954] env[65107]: DEBUG nova.network.neutron [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 786.870761] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6faaf5-3aa2-426f-923d-521416c45265 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.880128] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b431b7d3-6e87-4133-bc5d-bb2954f12f6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.915869] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ad0ce3-8174-4fca-b585-7922271ae933 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.924581] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c430b2-7efd-4484-ab4b-aef666f25d09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.942153] env[65107]: DEBUG nova.compute.provider_tree [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.978051] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc005476-6398-4865-9cfe-a4ddd633b93f tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "65d249a9-f79c-46c0-8630-169937cbcaf7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.747s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.080120] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.080334] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lock "8306ff90-b8bd-4270-8133-96abe483156b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.185044] env[65107]: WARNING neutronclient.v2_0.client [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.185971] env[65107]: WARNING openstack [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.186594] env[65107]: WARNING openstack [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.194320] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8d9ca07-f1dd-40cb-a684-c8a5994c49cb req-ad23cc76-4c59-4cfc-812f-c6ef5172b707 service nova] Releasing lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.248204] env[65107]: DEBUG nova.compute.manager [req-6b453bd9-3582-49c7-98ba-38fd92b856ee req-e0bd94ec-1e2e-43e8-bab3-47a6ca5f3bad service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Received event network-vif-plugged-a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 787.248439] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b453bd9-3582-49c7-98ba-38fd92b856ee req-e0bd94ec-1e2e-43e8-bab3-47a6ca5f3bad service nova] Acquiring lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.248639] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b453bd9-3582-49c7-98ba-38fd92b856ee req-e0bd94ec-1e2e-43e8-bab3-47a6ca5f3bad service nova] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.248803] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b453bd9-3582-49c7-98ba-38fd92b856ee req-e0bd94ec-1e2e-43e8-bab3-47a6ca5f3bad service nova] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.248971] env[65107]: DEBUG nova.compute.manager [req-6b453bd9-3582-49c7-98ba-38fd92b856ee req-e0bd94ec-1e2e-43e8-bab3-47a6ca5f3bad service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] No waiting events found dispatching network-vif-plugged-a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 787.249153] env[65107]: WARNING nova.compute.manager [req-6b453bd9-3582-49c7-98ba-38fd92b856ee req-e0bd94ec-1e2e-43e8-bab3-47a6ca5f3bad service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Received unexpected event network-vif-plugged-a8953e74-c080-4161-885f-b88b84a8bab2 for instance with vm_state building and task_state spawning. [ 787.300799] env[65107]: DEBUG nova.network.neutron [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.325234] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.325675] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.351053] env[65107]: WARNING openstack [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.351626] env[65107]: WARNING openstack [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.407622] env[65107]: DEBUG nova.network.neutron [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 787.443224] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.443608] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.471286] env[65107]: ERROR nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] [req-fbf5077e-3c76-4c92-b5b4-f9d9a39d8cfe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fbf5077e-3c76-4c92-b5b4-f9d9a39d8cfe"}]} [ 787.491477] env[65107]: DEBUG nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 787.507985] env[65107]: DEBUG nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 787.508301] env[65107]: DEBUG nova.compute.provider_tree [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 787.511622] env[65107]: WARNING neutronclient.v2_0.client [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.512292] env[65107]: WARNING openstack [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.512666] env[65107]: WARNING openstack [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.531186] env[65107]: DEBUG nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 787.554533] env[65107]: DEBUG nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 787.586384] env[65107]: DEBUG nova.compute.utils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 787.628821] env[65107]: WARNING neutronclient.v2_0.client [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.628821] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.629266] env[65107]: WARNING openstack [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.706481] env[65107]: DEBUG nova.network.neutron [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updated VIF entry in instance network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 787.706481] env[65107]: DEBUG nova.network.neutron [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.810472] env[65107]: INFO nova.compute.manager [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Took 2.82 seconds to deallocate network for instance. [ 788.089626] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lock "8306ff90-b8bd-4270-8133-96abe483156b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.092853] env[65107]: DEBUG nova.network.neutron [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Updating instance_info_cache with network_info: [{"id": "a8953e74-c080-4161-885f-b88b84a8bab2", "address": "fa:16:3e:41:7a:a4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8953e74-c0", "ovs_interfaceid": "a8953e74-c080-4161-885f-b88b84a8bab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 788.130180] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86994bcd-e79d-432b-90a0-85fcbe1a18e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.140571] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab51671-1a00-4951-a6a8-e50a4fb6deae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.177377] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a33d28-5c1b-4081-8914-5e9bf96049f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.186532] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21122812-a839-4417-924f-32d0d9ebeead {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.202345] env[65107]: DEBUG nova.compute.provider_tree [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 788.212518] env[65107]: DEBUG oslo_concurrency.lockutils [req-f40a16fd-368b-49a1-8684-45a428dbf333 req-d6bb49d1-b135-40da-a78a-737715b9631c service nova] Releasing lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 788.321171] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.597129] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Releasing lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 788.598064] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Instance network_info: |[{"id": "a8953e74-c080-4161-885f-b88b84a8bab2", "address": "fa:16:3e:41:7a:a4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8953e74-c0", "ovs_interfaceid": "a8953e74-c080-4161-885f-b88b84a8bab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 788.598805] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:7a:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8953e74-c080-4161-885f-b88b84a8bab2', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.608168] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 788.608766] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.608995] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-770839eb-70a3-4770-9a60-335456cbd4e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.635703] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.635703] env[65107]: value = "task-5102696" [ 788.635703] env[65107]: _type = "Task" [ 788.635703] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.646581] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102696, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.753799] env[65107]: DEBUG nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 71 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 788.754179] env[65107]: DEBUG nova.compute.provider_tree [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 71 to 72 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 788.754454] env[65107]: DEBUG nova.compute.provider_tree [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.100266] env[65107]: INFO nova.compute.manager [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Rebuilding instance [ 789.146810] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102696, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.156034] env[65107]: DEBUG nova.compute.manager [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 789.156155] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07f4cb6-4d7d-426c-87f0-6e0ff7978a63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.186217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.186217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lock "8306ff90-b8bd-4270-8133-96abe483156b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.186217] env[65107]: INFO nova.compute.manager [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Attaching volume 679309d6-17cf-4ee2-8196-c9517f0df270 to /dev/sdb [ 789.231174] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95de958d-cf13-4ade-b1e2-f2e8910c17f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.249308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5078bb8-fd05-428d-889a-804e54f4be2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.260851] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.529s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.268504] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.862s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.268815] env[65107]: DEBUG nova.objects.instance [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lazy-loading 'resources' on Instance uuid c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.270431] env[65107]: DEBUG nova.virt.block_device [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updating existing volume attachment record: 85dc6236-a353-41f4-9262-81bbabaf4d19 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 789.316905] env[65107]: INFO nova.scheduler.client.report [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Deleted allocations for instance d40aec5a-d898-42f3-a121-503cf043f72b [ 789.366268] env[65107]: DEBUG nova.compute.manager [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Received event network-vif-deleted-321da8b7-0ae1-438d-816f-6c711ef9b003 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 789.368471] env[65107]: DEBUG nova.compute.manager [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Received event network-vif-deleted-257c9dd4-8ac7-4bca-abb6-a32d3652c6a2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 789.368777] env[65107]: DEBUG nova.compute.manager [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 789.368952] env[65107]: DEBUG nova.compute.manager [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing instance network info cache due to event network-changed-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 789.369198] env[65107]: DEBUG oslo_concurrency.lockutils [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Acquiring lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.369365] env[65107]: DEBUG oslo_concurrency.lockutils [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Acquired lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.369491] env[65107]: DEBUG nova.network.neutron [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Refreshing network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 789.431353] env[65107]: DEBUG nova.compute.manager [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Received event network-changed-a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 789.431625] env[65107]: DEBUG nova.compute.manager [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Refreshing instance network info cache due to event network-changed-a8953e74-c080-4161-885f-b88b84a8bab2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 789.431847] env[65107]: DEBUG oslo_concurrency.lockutils [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Acquiring lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.431990] env[65107]: DEBUG oslo_concurrency.lockutils [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Acquired lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.432248] env[65107]: DEBUG nova.network.neutron [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Refreshing network info cache for port a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 789.647441] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102696, 'name': CreateVM_Task, 'duration_secs': 0.588082} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.647699] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.648249] env[65107]: WARNING neutronclient.v2_0.client [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.648670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.648825] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.649164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 789.649431] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62e1420c-adda-4ab7-82ce-e86a4e620aeb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.654694] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 789.654694] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522074b9-e707-466b-796b-af17798fbc92" [ 789.654694] env[65107]: _type = "Task" [ 789.654694] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.663978] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522074b9-e707-466b-796b-af17798fbc92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.831114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.831790] env[65107]: DEBUG oslo_concurrency.lockutils [None req-011f6f04-163d-4e96-91c5-d8fde1f57adb tempest-AttachInterfacesV270Test-1992497490 tempest-AttachInterfacesV270Test-1992497490-project-member] Lock "d40aec5a-d898-42f3-a121-503cf043f72b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.913s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.875083] env[65107]: WARNING neutronclient.v2_0.client [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.875746] env[65107]: WARNING openstack [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.876759] env[65107]: WARNING openstack [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.936314] env[65107]: WARNING neutronclient.v2_0.client [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.937073] env[65107]: WARNING openstack [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.937751] env[65107]: WARNING openstack [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.988451] env[65107]: WARNING openstack [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.988829] env[65107]: WARNING openstack [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.078953] env[65107]: WARNING neutronclient.v2_0.client [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.081415] env[65107]: WARNING openstack [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.081415] env[65107]: WARNING openstack [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.099045] env[65107]: WARNING openstack [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.099577] env[65107]: WARNING openstack [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.169675] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522074b9-e707-466b-796b-af17798fbc92, 'name': SearchDatastore_Task, 'duration_secs': 0.028796} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.171020] env[65107]: WARNING neutronclient.v2_0.client [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.171645] env[65107]: WARNING openstack [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.171988] env[65107]: WARNING openstack [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.180094] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 790.180793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.181062] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.181757] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.181757] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.181866] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.184932] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e80fd10-9485-4081-90b8-4a108f284d2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.186691] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d445f0a-ac84-4b4f-b9c1-ea019f9c2a35 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.205155] env[65107]: DEBUG nova.network.neutron [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updated VIF entry in instance network info cache for port 2ec2fffa-a684-4c3d-8ead-2c8b74d6252d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 790.205504] env[65107]: DEBUG nova.network.neutron [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [{"id": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "address": "fa:16:3e:50:ad:9d", "network": {"id": "d53d463e-8e2d-41d1-a9a0-558631bcf546", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1416727474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58588fe2a11f4ee0aab313903b211fa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec2fffa-a6", "ovs_interfaceid": "2ec2fffa-a684-4c3d-8ead-2c8b74d6252d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.210749] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 790.210749] env[65107]: value = "task-5102700" [ 790.210749] env[65107]: _type = "Task" [ 790.210749] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.211062] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.211243] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.216111] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a15cd90-717c-452b-892c-0ac0c39682dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.233016] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 790.233016] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5291d996-b4da-daf4-1936-5ae5ae4244c0" [ 790.233016] env[65107]: _type = "Task" [ 790.233016] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.233183] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.248332] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5291d996-b4da-daf4-1936-5ae5ae4244c0, 'name': SearchDatastore_Task, 'duration_secs': 0.015228} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.249915] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-843923e7-2153-4f3a-9763-a4ab544cfb41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.260095] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 790.260095] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5245e304-4b23-f963-129e-53391d0959bf" [ 790.260095] env[65107]: _type = "Task" [ 790.260095] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.276034] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5245e304-4b23-f963-129e-53391d0959bf, 'name': SearchDatastore_Task, 'duration_secs': 0.015714} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.276292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.276666] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090/a4f43fbc-63e3-4b69-bdb1-7aaf47aae090.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.277015] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca6564c8-cddf-46b9-b415-bb57cf0e4ccc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.286418] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 790.286418] env[65107]: value = "task-5102701" [ 790.286418] env[65107]: _type = "Task" [ 790.286418] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.297103] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.312618] env[65107]: DEBUG nova.network.neutron [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Updated VIF entry in instance network info cache for port a8953e74-c080-4161-885f-b88b84a8bab2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 790.313161] env[65107]: DEBUG nova.network.neutron [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Updating instance_info_cache with network_info: [{"id": "a8953e74-c080-4161-885f-b88b84a8bab2", "address": "fa:16:3e:41:7a:a4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8953e74-c0", "ovs_interfaceid": "a8953e74-c080-4161-885f-b88b84a8bab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.369877] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673b65fb-d046-427c-a731-41a4d1425cde {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.379851] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1547e9e-8c34-48a6-8231-2833bfe20e97 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.419594] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339ec039-c3e4-4a05-a336-45bf8b60b9d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.431722] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246aff23-9c49-4691-84e3-1fe7c2795f76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.453392] env[65107]: DEBUG nova.compute.provider_tree [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.506760] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "805e2210-fb58-48da-be63-5d548b2f8a2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 790.506993] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.700292] env[65107]: DEBUG nova.compute.manager [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 790.701267] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8729f7a-57f6-4eb9-a6a0-cc64db2f5296 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.712758] env[65107]: DEBUG oslo_concurrency.lockutils [req-35c740d6-8755-401a-baf8-3ef617b9a6fc req-ece1450c-43d2-4d39-beaa-a3e5b04701e2 service nova] Releasing lock "refresh_cache-ef417f32-8c0e-4a62-9078-cb2bf95b85f8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.726471] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102700, 'name': PowerOffVM_Task, 'duration_secs': 0.236655} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.726891] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 790.727193] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 790.728143] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a5274a-a99a-4f2a-9c77-ed85e60ffe30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.737190] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 790.737588] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-621f4872-fcf9-4f8d-b8a8-6a855ca1e7cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.770782] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 790.774734] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 790.774734] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Deleting the datastore file [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.774734] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f87b19b7-25a0-4399-af34-efd07729dfba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.783950] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 790.783950] env[65107]: value = "task-5102703" [ 790.783950] env[65107]: _type = "Task" [ 790.783950] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.797854] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102701, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.801427] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.816515] env[65107]: DEBUG oslo_concurrency.lockutils [req-7eca019c-9d9e-4af8-9f2d-a3de00b2d8f0 req-0c5433c9-ff22-4ab6-8ac7-9fb15825813e service nova] Releasing lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.958097] env[65107]: DEBUG nova.scheduler.client.report [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.009723] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 791.219026] env[65107]: INFO nova.compute.manager [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] instance snapshotting [ 791.219026] env[65107]: DEBUG nova.objects.instance [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'flavor' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 791.299247] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204258} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.303196] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 791.303384] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 791.303598] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.306529] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597132} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.307838] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090/a4f43fbc-63e3-4b69-bdb1-7aaf47aae090.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.307838] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.307838] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05c9e57e-646e-4746-a953-83d34275221b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.317639] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 791.317639] env[65107]: value = "task-5102704" [ 791.317639] env[65107]: _type = "Task" [ 791.317639] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.328900] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.464989] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.196s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.467817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 43.004s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.468081] env[65107]: DEBUG nova.objects.instance [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'resources' on Instance uuid c7739dfd-88d6-46ac-b3a0-e7dfe9b00332 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 791.511922] env[65107]: INFO nova.scheduler.client.report [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Deleted allocations for instance c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d [ 791.538507] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.724590] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e7aed6-0d55-4630-a351-c50aaa1a30c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.751380] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4834b1-fa4f-46cc-9310-262f1ea6b7ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.836993] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080459} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.837552] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.839131] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e802df8-50ee-48aa-98c5-b1c6a591f164 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.873086] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090/a4f43fbc-63e3-4b69-bdb1-7aaf47aae090.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.873827] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1dcbce8-4725-471e-a5f8-146e1d7b155a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.897953] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 791.897953] env[65107]: value = "task-5102706" [ 791.897953] env[65107]: _type = "Task" [ 791.897953] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.909749] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102706, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.025384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7ca082b7-9a58-46cd-936c-dbdb7b129c7a tempest-ServersAaction247Test-2109525583 tempest-ServersAaction247Test-2109525583-project-member] Lock "c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.888s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.072131] env[65107]: DEBUG nova.objects.instance [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lazy-loading 'flavor' on Instance uuid 1253473b-d4bc-49ac-9724-9aa1e7a8f038 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.266267] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 792.266267] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-89998522-820a-4d7c-bc16-a3338059dc3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.276056] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 792.276056] env[65107]: value = "task-5102707" [ 792.276056] env[65107]: _type = "Task" [ 792.276056] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.290431] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102707, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.371812] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 792.372080] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 792.372241] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 792.372423] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 792.372565] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 792.372706] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 792.372908] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.376523] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 792.376523] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 792.376523] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 792.376523] env[65107]: DEBUG nova.virt.hardware [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 792.376523] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c5449e-3f06-4bf6-8b53-13d2b46ba858 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.387146] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691822ef-60b3-4e21-8950-7ff7880ebc10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.409383] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.418471] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 792.425869] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.428691] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc5fd494-1e83-4ecd-8771-3af3ae132f58 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.451784] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102706, 'name': ReconfigVM_Task, 'duration_secs': 0.428986} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.457315] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522ad92e-31de-403d-5636-0663b99a8799/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 792.457315] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Reconfigured VM instance instance-00000030 to attach disk [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090/a4f43fbc-63e3-4b69-bdb1-7aaf47aae090.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.457315] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=65107) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 792.457315] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.457315] env[65107]: value = "task-5102708" [ 792.457315] env[65107]: _type = "Task" [ 792.457315] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.457964] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb260f49-2dec-4a7a-a4b5-f78ce5e84282 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.461186] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-519a39dc-8f68-4901-a373-1fba5059472a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.477433] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522ad92e-31de-403d-5636-0663b99a8799/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 792.477433] env[65107]: ERROR oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522ad92e-31de-403d-5636-0663b99a8799/disk-0.vmdk due to incomplete transfer. [ 792.483508] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1bc15670-ca43-4f2d-bba6-b2e4cda0c68b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.488155] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 792.488155] env[65107]: value = "task-5102709" [ 792.488155] env[65107]: _type = "Task" [ 792.488155] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.488155] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102708, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.497577] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522ad92e-31de-403d-5636-0663b99a8799/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 792.497577] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Uploaded image cbb434d5-e178-436d-a497-b20fc87120c6 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 792.500853] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 792.503938] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6c9c147e-2c9f-4571-be19-3e8cd903997d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.505227] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102709, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.515893] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 792.515893] env[65107]: value = "task-5102710" [ 792.515893] env[65107]: _type = "Task" [ 792.515893] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.525291] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102710, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.581435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.581435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.581843] env[65107]: WARNING neutronclient.v2_0.client [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.582579] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.582946] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.605248] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0752c715-867d-4fa4-a8b8-4d687843c33d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.615984] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5108b731-4e98-4497-bf17-d708ddedbfad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.662793] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bcef8f-154b-4faa-910f-a8b4efe0a2c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.672742] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c349d023-9c7f-4348-ae46-1a7dea271160 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.691359] env[65107]: DEBUG nova.compute.provider_tree [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.790043] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102707, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.975754] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102708, 'name': CreateVM_Task, 'duration_secs': 0.335423} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.975754] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 792.975754] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.975754] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.976544] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 792.976902] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40cd0a64-3ad6-4e06-b18d-b41a1bd8d6a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.983111] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 792.983111] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522c6813-a71f-b4f8-0272-db2b581a816a" [ 792.983111] env[65107]: _type = "Task" [ 792.983111] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.001060] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522c6813-a71f-b4f8-0272-db2b581a816a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.004776] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102709, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.080142} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.005509] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=65107) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 793.006454] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e754bd-0f2d-4a8c-be00-3968d71c763c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.037329] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090/ephemeral_0.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.041524] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bb03825-014d-4caa-8277-dda519693354 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.061630] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102710, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.063479] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 793.063479] env[65107]: value = "task-5102711" [ 793.063479] env[65107]: _type = "Task" [ 793.063479] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.073740] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102711, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.200845] env[65107]: DEBUG nova.scheduler.client.report [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 793.217135] env[65107]: DEBUG nova.network.neutron [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 793.245103] env[65107]: DEBUG nova.compute.manager [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 793.245103] env[65107]: DEBUG nova.compute.manager [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing instance network info cache due to event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 793.245601] env[65107]: DEBUG oslo_concurrency.lockutils [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.290291] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102707, 'name': CreateSnapshot_Task, 'duration_secs': 0.553999} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.290291] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 793.290291] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b14f280-fae1-495b-a54b-cb77649842bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.463164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.463553] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.464302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.464302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.464302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.467666] env[65107]: INFO nova.compute.manager [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Terminating instance [ 793.495728] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522c6813-a71f-b4f8-0272-db2b581a816a, 'name': SearchDatastore_Task, 'duration_secs': 0.016873} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.496109] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.496497] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.496775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.496928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.497234] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.497893] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92b78432-363d-4d04-be08-b5b5cc4364bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.509507] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.509825] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.510647] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72051145-d87d-4df4-9b2d-8811c8df6e56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.517442] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 793.517442] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52591f34-18e5-0d5f-e3fb-0c5ce20abbc6" [ 793.517442] env[65107]: _type = "Task" [ 793.517442] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.527402] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52591f34-18e5-0d5f-e3fb-0c5ce20abbc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.537884] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102710, 'name': Destroy_Task, 'duration_secs': 0.586512} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.538165] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Destroyed the VM [ 793.538498] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 793.538696] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c1c0a7dc-1279-49eb-b3cc-3e800bc41ffc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.546497] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 793.546497] env[65107]: value = "task-5102712" [ 793.546497] env[65107]: _type = "Task" [ 793.546497] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.556399] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102712, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.575319] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102711, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.709800] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.241s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.713572] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 43.242s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.719491] env[65107]: WARNING neutronclient.v2_0.client [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 793.720578] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.720890] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.737482] env[65107]: INFO nova.scheduler.client.report [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocations for instance c7739dfd-88d6-46ac-b3a0-e7dfe9b00332 [ 793.816924] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 793.816924] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b5ae7a4f-125d-4890-9f05-4461454318f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.830477] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 793.830477] env[65107]: value = "task-5102713" [ 793.830477] env[65107]: _type = "Task" [ 793.830477] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.843557] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102713, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.863947] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 793.865052] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992717', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'name': 'volume-679309d6-17cf-4ee2-8196-c9517f0df270', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8306ff90-b8bd-4270-8133-96abe483156b', 'attached_at': '', 'detached_at': '', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'serial': '679309d6-17cf-4ee2-8196-c9517f0df270'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 793.865306] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1662cd2-de5f-43c7-8508-35fa7f51848b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.890529] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5c34cb-61a7-4d97-b656-5ec8a20c9d18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.922451] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] volume-679309d6-17cf-4ee2-8196-c9517f0df270/volume-679309d6-17cf-4ee2-8196-c9517f0df270.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.922809] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-231ebf33-99f6-4fee-b23f-68631e9a359c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.938499] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.938882] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.953987] env[65107]: DEBUG oslo_vmware.api [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Waiting for the task: (returnval){ [ 793.953987] env[65107]: value = "task-5102714" [ 793.953987] env[65107]: _type = "Task" [ 793.953987] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.965138] env[65107]: DEBUG oslo_vmware.api [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102714, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.972776] env[65107]: DEBUG nova.compute.manager [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 793.972776] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.973547] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ed7067-0113-4aab-84a6-167d4e53de3a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.983653] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 793.983962] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-203c2886-e632-4514-8f51-d2fa842c079b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.031568] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52591f34-18e5-0d5f-e3fb-0c5ce20abbc6, 'name': SearchDatastore_Task, 'duration_secs': 0.013224} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.032144] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1602839-35b9-4b40-93c8-a2b8da87017a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.039379] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 794.039379] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202c1de-dfed-3fea-3a3e-90138b88fe94" [ 794.039379] env[65107]: _type = "Task" [ 794.039379] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.049834] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202c1de-dfed-3fea-3a3e-90138b88fe94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.063544] env[65107]: DEBUG oslo_vmware.api [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102712, 'name': RemoveSnapshot_Task, 'duration_secs': 0.362069} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.067058] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 794.067058] env[65107]: INFO nova.compute.manager [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Took 16.22 seconds to snapshot the instance on the hypervisor. [ 794.069436] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 794.070070] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 794.070172] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleting the datastore file [datastore2] 19518d90-ba6f-44e5-b7c7-1ececad2b54d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.077707] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f896aad8-52fc-4345-9390-edd40e16aef2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.089489] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102711, 'name': ReconfigVM_Task, 'duration_secs': 0.526887} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.091066] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Reconfigured VM instance instance-00000030 to attach disk [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090/ephemeral_0.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.092247] env[65107]: DEBUG oslo_vmware.api [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 794.092247] env[65107]: value = "task-5102716" [ 794.092247] env[65107]: _type = "Task" [ 794.092247] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.092486] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75dbe4c2-42a9-4c2b-8269-4836b5e54ebb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.098033] env[65107]: WARNING neutronclient.v2_0.client [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.098033] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.098033] env[65107]: WARNING openstack [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.118380] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 794.118380] env[65107]: value = "task-5102717" [ 794.118380] env[65107]: _type = "Task" [ 794.118380] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.131703] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102717, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.220411] env[65107]: INFO nova.compute.claims [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.249234] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bcce7abd-ebc6-4c3e-9258-9e8e7de0eee0 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "c7739dfd-88d6-46ac-b3a0-e7dfe9b00332" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 51.783s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.275623] env[65107]: DEBUG nova.network.neutron [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 794.279789] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 794.279789] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.343713] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102713, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.465403] env[65107]: DEBUG oslo_vmware.api [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102714, 'name': ReconfigVM_Task, 'duration_secs': 0.436832} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.465403] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Reconfigured VM instance instance-00000008 to attach disk [datastore2] volume-679309d6-17cf-4ee2-8196-c9517f0df270/volume-679309d6-17cf-4ee2-8196-c9517f0df270.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.470115] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb9f4fcc-ae87-46ba-bbcb-4ccbcb74b334 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.490035] env[65107]: DEBUG oslo_vmware.api [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Waiting for the task: (returnval){ [ 794.490035] env[65107]: value = "task-5102718" [ 794.490035] env[65107]: _type = "Task" [ 794.490035] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.505331] env[65107]: DEBUG oslo_vmware.api [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102718, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.551366] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202c1de-dfed-3fea-3a3e-90138b88fe94, 'name': SearchDatastore_Task, 'duration_secs': 0.027601} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.551637] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.551915] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.552275] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3092b09f-9e95-44b4-8026-45d3266c060c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.559935] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 794.559935] env[65107]: value = "task-5102719" [ 794.559935] env[65107]: _type = "Task" [ 794.559935] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.569348] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102719, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.572030] env[65107]: DEBUG nova.compute.manager [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Instance disappeared during snapshot {{(pid=65107) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 794.586050] env[65107]: DEBUG nova.compute.manager [None req-5b56056b-8ad9-4c01-96e2-ecb8bc0659f5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image not found during clean up cbb434d5-e178-436d-a497-b20fc87120c6 {{(pid=65107) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4658}} [ 794.606056] env[65107]: DEBUG oslo_vmware.api [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200043} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.606056] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.606056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.606257] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.606341] env[65107]: INFO nova.compute.manager [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Took 0.63 seconds to destroy the instance on the hypervisor. [ 794.606590] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 794.606785] env[65107]: DEBUG nova.compute.manager [-] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 794.606882] env[65107]: DEBUG nova.network.neutron [-] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 794.607135] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.607755] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.608106] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.628717] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102717, 'name': Rename_Task, 'duration_secs': 0.202466} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.629024] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.629336] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e763c6f-6e37-4408-9094-6b812876e488 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.637341] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 794.637341] env[65107]: value = "task-5102720" [ 794.637341] env[65107]: _type = "Task" [ 794.637341] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.649141] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102720, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.686736] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.728222] env[65107]: INFO nova.compute.resource_tracker [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating resource usage from migration 9e867391-8bde-4c95-983e-919c66e15cd7 [ 794.781381] env[65107]: DEBUG oslo_concurrency.lockutils [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.781876] env[65107]: DEBUG nova.compute.manager [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Inject network info {{(pid=65107) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 794.782247] env[65107]: DEBUG nova.compute.manager [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] network_info to inject: |[{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 794.788769] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Reconfiguring VM instance to set the machine id {{(pid=65107) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 794.793107] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 794.799822] env[65107]: DEBUG oslo_concurrency.lockutils [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.800043] env[65107]: DEBUG nova.network.neutron [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 794.801509] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ca67299-5684-44d4-8f16-3f1944319095 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.833282] env[65107]: DEBUG oslo_vmware.api [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 794.833282] env[65107]: value = "task-5102721" [ 794.833282] env[65107]: _type = "Task" [ 794.833282] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.854249] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102713, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.865993] env[65107]: DEBUG oslo_vmware.api [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102721, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.007096] env[65107]: DEBUG oslo_vmware.api [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102718, 'name': ReconfigVM_Task, 'duration_secs': 0.179309} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.007803] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992717', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'name': 'volume-679309d6-17cf-4ee2-8196-c9517f0df270', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8306ff90-b8bd-4270-8133-96abe483156b', 'attached_at': '', 'detached_at': '', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'serial': '679309d6-17cf-4ee2-8196-c9517f0df270'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 795.081765] env[65107]: DEBUG nova.objects.instance [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lazy-loading 'flavor' on Instance uuid 1253473b-d4bc-49ac-9724-9aa1e7a8f038 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.092095] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102719, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.148956] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102720, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.320716] env[65107]: WARNING neutronclient.v2_0.client [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.321426] env[65107]: WARNING openstack [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.321786] env[65107]: WARNING openstack [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.351081] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102713, 'name': CloneVM_Task, 'duration_secs': 1.423082} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.357522] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created linked-clone VM from snapshot [ 795.358358] env[65107]: DEBUG oslo_vmware.api [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102721, 'name': ReconfigVM_Task, 'duration_secs': 0.219065} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.359980] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.360936] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c506e50-2fb5-450f-a810-a85ba6ed28d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.363809] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-983ceb1a-5f7b-4db5-9482-b8397ea1e363 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Reconfigured VM instance to set the machine id {{(pid=65107) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 795.374022] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Uploading image 7cb99544-c916-4db5-b034-88926221b037 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 795.420083] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 795.420083] env[65107]: value = "vm-992720" [ 795.420083] env[65107]: _type = "VirtualMachine" [ 795.420083] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 795.420612] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "40474981-eeef-492e-8a8b-aaea7f554c02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.420928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.423021] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e566ec79-0397-49cb-a295-33d32b960da9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.442840] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease: (returnval){ [ 795.442840] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e6433-5670-3c1f-f006-edaf3956327e" [ 795.442840] env[65107]: _type = "HttpNfcLease" [ 795.442840] env[65107]: } obtained for exporting VM: (result){ [ 795.442840] env[65107]: value = "vm-992720" [ 795.442840] env[65107]: _type = "VirtualMachine" [ 795.442840] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 795.443261] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the lease: (returnval){ [ 795.443261] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e6433-5670-3c1f-f006-edaf3956327e" [ 795.443261] env[65107]: _type = "HttpNfcLease" [ 795.443261] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 795.445756] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57405513-a4f3-4f69-94ab-1f3d09a93c3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.459148] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 795.459148] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e6433-5670-3c1f-f006-edaf3956327e" [ 795.459148] env[65107]: _type = "HttpNfcLease" [ 795.459148] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 795.459883] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 795.459883] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e6433-5670-3c1f-f006-edaf3956327e" [ 795.459883] env[65107]: _type = "HttpNfcLease" [ 795.459883] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 795.460471] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e67f529-fbdb-4e0d-810e-044b2e9e6c5b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.464247] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055f9c41-5944-4a2f-aacc-76556f291383 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.472049] env[65107]: DEBUG nova.network.neutron [-] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.504258] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52329414-72f5-c2ce-4576-d1f41f55ecf3/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 795.504464] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52329414-72f5-c2ce-4576-d1f41f55ecf3/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 795.506411] env[65107]: INFO nova.compute.manager [-] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Took 0.90 seconds to deallocate network for instance. [ 795.507376] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c109f004-74de-487f-b2f9-3b51928bcbaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.580941] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.588620] env[65107]: WARNING openstack [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.588620] env[65107]: WARNING openstack [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.600957] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4e76f30-750a-4442-b728-b75367b4d929 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "080275a2-6b23-4344-a2d5-cde1fc1a4909" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.601185] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4e76f30-750a-4442-b728-b75367b4d929 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "080275a2-6b23-4344-a2d5-cde1fc1a4909" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.603355] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb75612f-33af-4fde-bfaf-f61246339e07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.609674] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.615966] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102719, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542321} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.624560] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.624824] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.626239] env[65107]: DEBUG nova.compute.provider_tree [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.627402] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-630f817d-e31a-48a9-a5b4-59c728c5f9f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.636693] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 795.636693] env[65107]: value = "task-5102723" [ 795.636693] env[65107]: _type = "Task" [ 795.636693] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.660470] env[65107]: DEBUG oslo_vmware.api [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102720, 'name': PowerOnVM_Task, 'duration_secs': 0.729556} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.660710] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102723, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.660957] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.661200] env[65107]: INFO nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Took 9.63 seconds to spawn the instance on the hypervisor. [ 795.661405] env[65107]: DEBUG nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 795.662455] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc43a74e-0894-424f-a57d-e6f960a43aaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.686758] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-41ffc327-127f-445a-b4bf-9508f7b55493 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.696117] env[65107]: WARNING neutronclient.v2_0.client [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.697285] env[65107]: WARNING openstack [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.697780] env[65107]: WARNING openstack [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.755236] env[65107]: DEBUG nova.compute.manager [req-6bbc88b4-359a-486b-8660-3cdd5d2fd2d2 req-6c2c5b63-3e74-43f8-bc88-9a4139d6e7d7 service nova] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Received event network-vif-deleted-a6da0470-a2c9-4d2d-869b-c102c75b6b54 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 795.792176] env[65107]: DEBUG nova.network.neutron [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updated VIF entry in instance network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 795.792565] env[65107]: DEBUG nova.network.neutron [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.924952] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 796.101138] env[65107]: DEBUG nova.objects.instance [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lazy-loading 'flavor' on Instance uuid 8306ff90-b8bd-4270-8133-96abe483156b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.109351] env[65107]: DEBUG nova.compute.manager [None req-b4e76f30-750a-4442-b728-b75367b4d929 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 080275a2-6b23-4344-a2d5-cde1fc1a4909] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 796.130861] env[65107]: DEBUG nova.scheduler.client.report [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 796.154344] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102723, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.27296} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.154985] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.155989] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801af3f0-dd1d-4627-9569-845edf8b3465 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.180807] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.186553] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b09bb764-cb2a-4587-985d-cf6c6520dd54 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.208403] env[65107]: INFO nova.compute.manager [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Took 53.10 seconds to build instance. [ 796.215598] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.215931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.217662] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 796.217662] env[65107]: value = "task-5102724" [ 796.217662] env[65107]: _type = "Task" [ 796.217662] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.231136] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102724, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.295986] env[65107]: DEBUG oslo_concurrency.lockutils [req-eba7b286-f9ff-49b1-afa3-548c9307797a req-7aebfc55-a178-484a-802c-c3d2fab997c0 service nova] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 796.296972] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 796.297537] env[65107]: WARNING neutronclient.v2_0.client [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 796.459034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.609098] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea81b19-b4d0-4f07-ae4e-82d6af52c353 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lock "8306ff90-b8bd-4270-8133-96abe483156b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.424s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.613065] env[65107]: DEBUG nova.compute.manager [None req-b4e76f30-750a-4442-b728-b75367b4d929 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 080275a2-6b23-4344-a2d5-cde1fc1a4909] Instance disappeared before build. {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2482}} [ 796.637389] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.926s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.637735] env[65107]: INFO nova.compute.manager [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Migrating [ 796.645775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 45.607s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.646133] env[65107]: DEBUG nova.objects.instance [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lazy-loading 'resources' on Instance uuid b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.715920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-96092635-cff1-4b0d-b886-ed106e642e59 tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.620s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.718846] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 796.738461] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102724, 'name': ReconfigVM_Task, 'duration_secs': 0.510477} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.738461] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7/65d249a9-f79c-46c0-8630-169937cbcaf7.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.738461] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be8b29e1-d31c-43b4-a443-a35ef7cbe8e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.741514] env[65107]: DEBUG nova.network.neutron [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 796.752851] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 796.752851] env[65107]: value = "task-5102725" [ 796.752851] env[65107]: _type = "Task" [ 796.752851] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.770665] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102725, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.123075] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "9a46ff51-4035-4709-ae6c-0fb5397200ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.123333] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.130062] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4e76f30-750a-4442-b728-b75367b4d929 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "080275a2-6b23-4344-a2d5-cde1fc1a4909" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.529s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.159134] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.159809] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.159809] env[65107]: DEBUG nova.network.neutron [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 797.248150] env[65107]: WARNING neutronclient.v2_0.client [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 797.248919] env[65107]: WARNING openstack [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.249287] env[65107]: WARNING openstack [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.260624] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.273593] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102725, 'name': Rename_Task, 'duration_secs': 0.189714} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.277458] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.278545] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7bf7820-e828-40f7-a1dc-b4d7d67ca319 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.286298] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 797.286298] env[65107]: value = "task-5102726" [ 797.286298] env[65107]: _type = "Task" [ 797.286298] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.299325] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102726, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.434400] env[65107]: WARNING openstack [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.434400] env[65107]: WARNING openstack [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.517493] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "701a1a16-9348-4748-bb24-ee245a76566c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.517894] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.586884] env[65107]: WARNING neutronclient.v2_0.client [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 797.587494] env[65107]: WARNING openstack [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.587881] env[65107]: WARNING openstack [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.628517] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 797.664875] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 797.665613] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.665982] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.699023] env[65107]: DEBUG nova.network.neutron [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 797.800812] env[65107]: DEBUG oslo_vmware.api [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102726, 'name': PowerOnVM_Task, 'duration_secs': 0.476049} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.802379] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.802850] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.811889] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 797.812292] env[65107]: DEBUG nova.compute.manager [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 797.813566] env[65107]: DEBUG nova.compute.manager [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 797.813756] env[65107]: DEBUG nova.compute.manager [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing instance network info cache due to event network-changed-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 797.813992] env[65107]: DEBUG oslo_concurrency.lockutils [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Acquiring lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.818081] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0218fe9d-38b6-4548-b540-8f1b214e5ca0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.834324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f0307f-38d6-43a7-bcdb-61f0f435f3cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.848051] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f083ba0a-18e5-40d8-9f7b-674f4395a976 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.889619] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 797.891637] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.891637] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.899871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745f9b3a-802c-494a-b0da-c8c957e3d39a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.910897] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d90695-1202-4d4a-9b1d-42aa28d75250 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.932759] env[65107]: DEBUG nova.compute.provider_tree [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.957209] env[65107]: DEBUG nova.compute.manager [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Received event network-changed-a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 797.957516] env[65107]: DEBUG nova.compute.manager [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Refreshing instance network info cache due to event network-changed-a8953e74-c080-4161-885f-b88b84a8bab2. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 797.957785] env[65107]: DEBUG oslo_concurrency.lockutils [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Acquiring lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.958013] env[65107]: DEBUG oslo_concurrency.lockutils [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Acquired lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.958827] env[65107]: DEBUG nova.network.neutron [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Refreshing network info cache for port a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 798.003088] env[65107]: DEBUG nova.network.neutron [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.022278] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 798.106233] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.106452] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lock "8306ff90-b8bd-4270-8133-96abe483156b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.153038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.202594] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.202594] env[65107]: DEBUG nova.compute.manager [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Inject network info {{(pid=65107) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 798.202594] env[65107]: DEBUG nova.compute.manager [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] network_info to inject: |[{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 798.208404] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Reconfiguring VM instance to set the machine id {{(pid=65107) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 798.208404] env[65107]: DEBUG oslo_concurrency.lockutils [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Acquired lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.208404] env[65107]: DEBUG nova.network.neutron [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Refreshing network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 798.209174] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec1641a2-66b7-42e4-adfd-e91c9db613b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.227933] env[65107]: DEBUG oslo_vmware.api [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 798.227933] env[65107]: value = "task-5102727" [ 798.227933] env[65107]: _type = "Task" [ 798.227933] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.238341] env[65107]: DEBUG oslo_vmware.api [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102727, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.337027] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.439301] env[65107]: DEBUG nova.scheduler.client.report [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 798.462291] env[65107]: WARNING neutronclient.v2_0.client [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.463082] env[65107]: WARNING openstack [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.463514] env[65107]: WARNING openstack [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.505832] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.542145] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.589418] env[65107]: WARNING openstack [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.589939] env[65107]: WARNING openstack [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.613600] env[65107]: INFO nova.compute.manager [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Detaching volume 679309d6-17cf-4ee2-8196-c9517f0df270 [ 798.653278] env[65107]: INFO nova.virt.block_device [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Attempting to driver detach volume 679309d6-17cf-4ee2-8196-c9517f0df270 from mountpoint /dev/sdb [ 798.653593] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 798.653851] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992717', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'name': 'volume-679309d6-17cf-4ee2-8196-c9517f0df270', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8306ff90-b8bd-4270-8133-96abe483156b', 'attached_at': '', 'detached_at': '', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'serial': '679309d6-17cf-4ee2-8196-c9517f0df270'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 798.655596] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea385b9-24ef-41d7-9aeb-cd7a8e48001b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.663112] env[65107]: WARNING neutronclient.v2_0.client [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.663927] env[65107]: WARNING openstack [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.664353] env[65107]: WARNING openstack [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.692607] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fcb411-7553-4f00-a6ab-447e113d788a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.702033] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff49c01-51f8-4989-b21e-72e83bedec27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.727109] env[65107]: WARNING neutronclient.v2_0.client [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.727901] env[65107]: WARNING openstack [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.728311] env[65107]: WARNING openstack [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.741313] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2045ad-dd3d-420b-880a-a84529d81dc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.776917] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] The volume has not been displaced from its original location: [datastore2] volume-679309d6-17cf-4ee2-8196-c9517f0df270/volume-679309d6-17cf-4ee2-8196-c9517f0df270.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 798.786871] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Reconfiguring VM instance instance-00000008 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 798.787550] env[65107]: DEBUG oslo_vmware.api [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102727, 'name': ReconfigVM_Task, 'duration_secs': 0.172799} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.791469] env[65107]: DEBUG nova.network.neutron [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Updated VIF entry in instance network info cache for port a8953e74-c080-4161-885f-b88b84a8bab2. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 798.791988] env[65107]: DEBUG nova.network.neutron [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Updating instance_info_cache with network_info: [{"id": "a8953e74-c080-4161-885f-b88b84a8bab2", "address": "fa:16:3e:41:7a:a4", "network": {"id": "c689f83b-feb1-455f-8498-c96c9ed4b43a", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-731299484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "634f89d91e1443c7a0eb1314ecb17232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8953e74-c0", "ovs_interfaceid": "a8953e74-c080-4161-885f-b88b84a8bab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.797251] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d98ebec9-7ce4-428f-8a05-12f20b62abbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.811805] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b1d8ae04-3882-4e6e-993f-0bf8b7b6ba27 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Reconfigured VM instance to set the machine id {{(pid=65107) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 798.822749] env[65107]: DEBUG oslo_vmware.api [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Waiting for the task: (returnval){ [ 798.822749] env[65107]: value = "task-5102728" [ 798.822749] env[65107]: _type = "Task" [ 798.822749] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.835341] env[65107]: DEBUG oslo_vmware.api [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102728, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.895584] env[65107]: WARNING openstack [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.896064] env[65107]: WARNING openstack [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.902892] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.903132] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.903361] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.903554] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.903751] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.906068] env[65107]: INFO nova.compute.manager [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Terminating instance [ 798.945606] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.300s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.948967] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.072s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.950799] env[65107]: INFO nova.compute.claims [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.977072] env[65107]: WARNING neutronclient.v2_0.client [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.977904] env[65107]: WARNING openstack [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.978406] env[65107]: WARNING openstack [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.989834] env[65107]: INFO nova.scheduler.client.report [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Deleted allocations for instance b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8 [ 799.109553] env[65107]: DEBUG nova.network.neutron [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updated VIF entry in instance network info cache for port fa708f72-f0b6-4ac5-a002-0d6589e8ab0e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 799.110130] env[65107]: DEBUG nova.network.neutron [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [{"id": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "address": "fa:16:3e:fa:7d:95", "network": {"id": "0541f99d-49a1-4dc8-b772-ec2b653853d4", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-2007985104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bcb6424259344f3fa87ba9a4c10637d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa708f72-f0", "ovs_interfaceid": "fa708f72-f0b6-4ac5-a002-0d6589e8ab0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 799.111920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "65d249a9-f79c-46c0-8630-169937cbcaf7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.112276] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "65d249a9-f79c-46c0-8630-169937cbcaf7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.112650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "65d249a9-f79c-46c0-8630-169937cbcaf7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.112911] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "65d249a9-f79c-46c0-8630-169937cbcaf7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.113123] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "65d249a9-f79c-46c0-8630-169937cbcaf7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.116093] env[65107]: INFO nova.compute.manager [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Terminating instance [ 799.315767] env[65107]: DEBUG oslo_concurrency.lockutils [req-ce2944eb-1fcf-46b2-9981-e8d80b8dfc2d req-6ed425b5-0dbf-430c-ac2b-1ca2d1367791 service nova] Releasing lock "refresh_cache-a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.333624] env[65107]: DEBUG oslo_vmware.api [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102728, 'name': ReconfigVM_Task, 'duration_secs': 0.342323} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.333996] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Reconfigured VM instance instance-00000008 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 799.339265] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92207354-433b-4aff-b1c5-73bec8754afd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.355332] env[65107]: DEBUG oslo_vmware.api [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Waiting for the task: (returnval){ [ 799.355332] env[65107]: value = "task-5102729" [ 799.355332] env[65107]: _type = "Task" [ 799.355332] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.368644] env[65107]: DEBUG oslo_vmware.api [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.409815] env[65107]: DEBUG nova.compute.manager [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 799.410099] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.411070] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7e1d71-f9ed-42d3-a395-cbf68d80c344 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.420613] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.420971] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d62cd4c-b9de-4fff-956f-7ee761b7108d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.429204] env[65107]: DEBUG oslo_vmware.api [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 799.429204] env[65107]: value = "task-5102730" [ 799.429204] env[65107]: _type = "Task" [ 799.429204] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.444813] env[65107]: DEBUG oslo_vmware.api [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.501041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfe6d078-eaaf-480c-a820-3542af5cfa69 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 52.464s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.616500] env[65107]: DEBUG oslo_concurrency.lockutils [req-b7ed8a2b-3edd-4b05-9755-48c1fc0e86d7 req-2c3be52c-51ea-4c57-9f7b-da53197db2e0 service nova] Releasing lock "refresh_cache-1253473b-d4bc-49ac-9724-9aa1e7a8f038" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.620409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "refresh_cache-65d249a9-f79c-46c0-8630-169937cbcaf7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.620601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquired lock "refresh_cache-65d249a9-f79c-46c0-8630-169937cbcaf7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.620787] env[65107]: DEBUG nova.network.neutron [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 799.866023] env[65107]: DEBUG oslo_vmware.api [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Task: {'id': task-5102729, 'name': ReconfigVM_Task, 'duration_secs': 0.191885} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.866389] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992717', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'name': 'volume-679309d6-17cf-4ee2-8196-c9517f0df270', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8306ff90-b8bd-4270-8133-96abe483156b', 'attached_at': '', 'detached_at': '', 'volume_id': '679309d6-17cf-4ee2-8196-c9517f0df270', 'serial': '679309d6-17cf-4ee2-8196-c9517f0df270'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 799.939848] env[65107]: DEBUG oslo_vmware.api [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102730, 'name': PowerOffVM_Task, 'duration_secs': 0.386945} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.940560] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 799.940560] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 799.940728] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7599132b-09f5-40e3-b49b-6a606aeb4198 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.020418] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 800.020639] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 800.020782] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Deleting the datastore file [datastore2] 1253473b-d4bc-49ac-9724-9aa1e7a8f038 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 800.021609] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24b1cc82-89a7-4007-9d4f-6008a1a0932e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.024538] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a403ad7-da90-46ae-8d1d-8b6fe4fbd329 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.035124] env[65107]: DEBUG oslo_vmware.api [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for the task: (returnval){ [ 800.035124] env[65107]: value = "task-5102732" [ 800.035124] env[65107]: _type = "Task" [ 800.035124] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.049837] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 800.066354] env[65107]: DEBUG oslo_vmware.api [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102732, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.124124] env[65107]: WARNING neutronclient.v2_0.client [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.125625] env[65107]: WARNING openstack [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.125625] env[65107]: WARNING openstack [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.155830] env[65107]: DEBUG nova.network.neutron [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 800.216518] env[65107]: DEBUG nova.network.neutron [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 800.396684] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.398309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.398309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.398309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.398309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.400121] env[65107]: INFO nova.compute.manager [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Terminating instance [ 800.436669] env[65107]: DEBUG nova.objects.instance [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lazy-loading 'flavor' on Instance uuid 8306ff90-b8bd-4270-8133-96abe483156b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.518450] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c7db32-2550-47be-ba4f-ec6f870f3cf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.527282] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f1516b-ec07-470c-908f-9257fc430be3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.563630] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 800.567220] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c5c62c2-1efb-4041-baa6-eea1a64ba950 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.569304] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcf3e97-5cec-48e2-94ea-1a539a34119b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.577843] env[65107]: DEBUG oslo_vmware.api [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Task: {'id': task-5102732, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257718} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.581403] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 800.581645] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 800.581828] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 800.581999] env[65107]: INFO nova.compute.manager [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Took 1.17 seconds to destroy the instance on the hypervisor. [ 800.582303] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 800.582751] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 800.582751] env[65107]: value = "task-5102733" [ 800.582751] env[65107]: _type = "Task" [ 800.582751] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.582984] env[65107]: DEBUG nova.compute.manager [-] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 800.583375] env[65107]: DEBUG nova.network.neutron [-] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 800.583453] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.583943] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.584284] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.592086] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e81cf2-535d-42d3-a568-7e6229bf706b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.607454] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.616981] env[65107]: DEBUG nova.compute.provider_tree [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.643032] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.721571] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Releasing lock "refresh_cache-65d249a9-f79c-46c0-8630-169937cbcaf7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 800.722092] env[65107]: DEBUG nova.compute.manager [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 800.722365] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.723381] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37494aa-9a5b-42cf-883e-9d7f56ff0ba4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.732472] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 800.732751] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dcdbf25-2608-465d-89f3-dbc8a5d6e92d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.740400] env[65107]: DEBUG oslo_vmware.api [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 800.740400] env[65107]: value = "task-5102734" [ 800.740400] env[65107]: _type = "Task" [ 800.740400] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.750358] env[65107]: DEBUG oslo_vmware.api [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.907967] env[65107]: DEBUG nova.compute.manager [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 800.907967] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.908988] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecc0c4c-758f-4b04-a9ce-1429c6866d1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.918288] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 800.919203] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83aa186c-5732-4345-8583-d3d3415225e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.926570] env[65107]: DEBUG oslo_vmware.api [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 800.926570] env[65107]: value = "task-5102735" [ 800.926570] env[65107]: _type = "Task" [ 800.926570] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.936886] env[65107]: DEBUG oslo_vmware.api [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.106478] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102733, 'name': PowerOffVM_Task, 'duration_secs': 0.268992} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.107240] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.107598] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 801.122024] env[65107]: DEBUG nova.scheduler.client.report [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 801.190250] env[65107]: DEBUG nova.compute.manager [req-114c2308-33d2-4fa6-adbc-27a0805f7f06 req-ea15ac2e-3e9e-4a44-820d-055f19c5fb54 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Received event network-vif-deleted-fa708f72-f0b6-4ac5-a002-0d6589e8ab0e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 801.190250] env[65107]: INFO nova.compute.manager [req-114c2308-33d2-4fa6-adbc-27a0805f7f06 req-ea15ac2e-3e9e-4a44-820d-055f19c5fb54 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Neutron deleted interface fa708f72-f0b6-4ac5-a002-0d6589e8ab0e; detaching it from the instance and deleting it from the info cache [ 801.190408] env[65107]: DEBUG nova.network.neutron [req-114c2308-33d2-4fa6-adbc-27a0805f7f06 req-ea15ac2e-3e9e-4a44-820d-055f19c5fb54 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 801.251801] env[65107]: DEBUG oslo_vmware.api [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102734, 'name': PowerOffVM_Task, 'duration_secs': 0.218019} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.252191] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.252358] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.252632] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2207d9d-855a-4a97-a1b8-28cb8039589f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.282697] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.283230] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.283230] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Deleting the datastore file [datastore1] 65d249a9-f79c-46c0-8630-169937cbcaf7 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.283412] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f7c5504-fd48-424c-8600-ee0bdbd06a4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.292043] env[65107]: DEBUG oslo_vmware.api [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for the task: (returnval){ [ 801.292043] env[65107]: value = "task-5102737" [ 801.292043] env[65107]: _type = "Task" [ 801.292043] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.300668] env[65107]: DEBUG oslo_vmware.api [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.438316] env[65107]: DEBUG oslo_vmware.api [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102735, 'name': PowerOffVM_Task, 'duration_secs': 0.214498} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.438660] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.438895] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.439392] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3b62a71-d62c-41b7-9bbe-360195b5a11a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.445586] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4ffd9ebb-ee02-4f6d-b25b-006f40af0226 tempest-VolumesAssistedSnapshotsTest-1944867076 tempest-VolumesAssistedSnapshotsTest-1944867076-project-admin] Lock "8306ff90-b8bd-4270-8133-96abe483156b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.338s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.519991] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.520175] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.520361] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Deleting the datastore file [datastore2] f2814dd6-8be5-49b2-a030-f57e452c2e6d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.520675] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42fe10c0-c8c2-41d1-922c-f51b193592f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.529024] env[65107]: DEBUG oslo_vmware.api [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for the task: (returnval){ [ 801.529024] env[65107]: value = "task-5102739" [ 801.529024] env[65107]: _type = "Task" [ 801.529024] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.538523] env[65107]: DEBUG oslo_vmware.api [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.615655] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 801.615992] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.616198] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 801.616387] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.616536] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 801.616713] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 801.616956] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.617138] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 801.617407] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 801.617470] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 801.617753] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 801.622882] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a21f23ff-06b7-41de-bfdd-93da1a7cd761 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.634548] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.635077] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 801.637963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.769s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.640022] env[65107]: INFO nova.compute.claims [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.649795] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 801.649795] env[65107]: value = "task-5102740" [ 801.649795] env[65107]: _type = "Task" [ 801.649795] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.660074] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102740, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.673674] env[65107]: DEBUG nova.network.neutron [-] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 801.693062] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-222e24c5-a125-4221-80cd-c0b95e75b90e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.703452] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852bb0ac-606d-476f-9d51-660bd9a35553 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.745307] env[65107]: DEBUG nova.compute.manager [req-114c2308-33d2-4fa6-adbc-27a0805f7f06 req-ea15ac2e-3e9e-4a44-820d-055f19c5fb54 service nova] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Detach interface failed, port_id=fa708f72-f0b6-4ac5-a002-0d6589e8ab0e, reason: Instance 1253473b-d4bc-49ac-9724-9aa1e7a8f038 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 801.803019] env[65107]: DEBUG oslo_vmware.api [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Task: {'id': task-5102737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188312} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.803350] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.803623] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 801.803908] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.804218] env[65107]: INFO nova.compute.manager [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Took 1.08 seconds to destroy the instance on the hypervisor. [ 801.804593] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 801.804801] env[65107]: DEBUG nova.compute.manager [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 801.804903] env[65107]: DEBUG nova.network.neutron [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 801.805211] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.805804] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.806115] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.832404] env[65107]: DEBUG nova.network.neutron [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 801.832684] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.040346] env[65107]: DEBUG oslo_vmware.api [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Task: {'id': task-5102739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294567} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.040630] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.040820] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 802.041008] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.041200] env[65107]: INFO nova.compute.manager [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 802.041455] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 802.041657] env[65107]: DEBUG nova.compute.manager [-] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 802.041760] env[65107]: DEBUG nova.network.neutron [-] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 802.042031] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.042579] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.042877] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.094291] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.147334] env[65107]: DEBUG nova.compute.utils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 802.149687] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 802.150172] env[65107]: DEBUG nova.network.neutron [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 802.151054] env[65107]: WARNING neutronclient.v2_0.client [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.151152] env[65107]: WARNING neutronclient.v2_0.client [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.152125] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.152646] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.175946] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102740, 'name': ReconfigVM_Task, 'duration_secs': 0.259956} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.176531] env[65107]: INFO nova.compute.manager [-] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Took 1.59 seconds to deallocate network for instance. [ 802.176947] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 802.225167] env[65107]: DEBUG nova.policy [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6723c29f05774e138ee11fa945c058e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '139a38d3b7bd4d3a8861d2e1e6c7a2c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 802.335685] env[65107]: DEBUG nova.network.neutron [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.546537] env[65107]: DEBUG nova.compute.manager [req-fdb1f66d-eef7-4924-98f7-0027f68c64d1 req-f8ecce72-d08f-4489-8d17-998a313833e9 service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Received event network-vif-deleted-0b52bca6-ff8c-4322-9b18-1c1129186df6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 802.546782] env[65107]: INFO nova.compute.manager [req-fdb1f66d-eef7-4924-98f7-0027f68c64d1 req-f8ecce72-d08f-4489-8d17-998a313833e9 service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Neutron deleted interface 0b52bca6-ff8c-4322-9b18-1c1129186df6; detaching it from the instance and deleting it from the info cache [ 802.546947] env[65107]: DEBUG nova.network.neutron [req-fdb1f66d-eef7-4924-98f7-0027f68c64d1 req-f8ecce72-d08f-4489-8d17-998a313833e9 service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.648064] env[65107]: DEBUG nova.network.neutron [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Successfully created port: cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 802.663594] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 802.694099] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:22:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a3c77cac-4e5e-476c-bee7-e5d40e2e12d5',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-991042522',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 802.694099] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 802.694099] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 802.694099] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 802.694099] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 802.694099] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 802.696069] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.696069] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 802.696069] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 802.696069] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 802.696069] env[65107]: DEBUG nova.virt.hardware [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 802.701849] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfiguring VM instance instance-00000027 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 802.703291] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.703903] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f7d8b7e-e02f-40c6-8260-f504ce4e4137 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.732217] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 802.732217] env[65107]: value = "task-5102742" [ 802.732217] env[65107]: _type = "Task" [ 802.732217] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.745019] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102742, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.839451] env[65107]: INFO nova.compute.manager [-] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Took 1.03 seconds to deallocate network for instance. [ 802.880196] env[65107]: DEBUG nova.network.neutron [-] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 803.051115] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0f280be-b096-4d5f-9500-62ee644b9056 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.068955] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70567481-8508-41c1-8728-a875988ec174 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.101656] env[65107]: DEBUG nova.compute.manager [req-fdb1f66d-eef7-4924-98f7-0027f68c64d1 req-f8ecce72-d08f-4489-8d17-998a313833e9 service nova] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Detach interface failed, port_id=0b52bca6-ff8c-4322-9b18-1c1129186df6, reason: Instance f2814dd6-8be5-49b2-a030-f57e452c2e6d could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 803.232875] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9575730a-4de2-4ce6-82ba-a9ebe1fd0bf4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.245543] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102742, 'name': ReconfigVM_Task, 'duration_secs': 0.187688} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.247608] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfigured VM instance instance-00000027 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 803.248480] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d989e9a3-fb8a-4597-bd33-76893d950987 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.251954] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7117a297-5ad2-4f84-b360-d4b2ce4db022 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.276509] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.302987] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24601429-94d7-473d-9f00-15730ea88c8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.318094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f774ca6-f3a1-40c1-9174-21e10ff05a0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.328414] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae03bd28-acd7-447f-a0f5-45906de04ff9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.333206] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 803.333206] env[65107]: value = "task-5102743" [ 803.333206] env[65107]: _type = "Task" [ 803.333206] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.345201] env[65107]: DEBUG nova.compute.provider_tree [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.348829] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.353803] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102743, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.383038] env[65107]: INFO nova.compute.manager [-] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Took 1.34 seconds to deallocate network for instance. [ 803.577437] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52329414-72f5-c2ce-4576-d1f41f55ecf3/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 803.578496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f7dc54-56b9-4588-b550-21a1408a738e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.586223] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52329414-72f5-c2ce-4576-d1f41f55ecf3/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 803.586471] env[65107]: ERROR oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52329414-72f5-c2ce-4576-d1f41f55ecf3/disk-0.vmdk due to incomplete transfer. [ 803.586744] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9e0c24c3-5d85-422c-94d8-3e81f445b067 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.597802] env[65107]: DEBUG oslo_vmware.rw_handles [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52329414-72f5-c2ce-4576-d1f41f55ecf3/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 803.598052] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Uploaded image 7cb99544-c916-4db5-b034-88926221b037 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 803.600831] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 803.601133] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8f85c401-31e7-46bf-8e9f-5dbfa3b005f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.608785] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 803.608785] env[65107]: value = "task-5102744" [ 803.608785] env[65107]: _type = "Task" [ 803.608785] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.619375] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102744, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.677497] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 803.702613] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 803.703049] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 803.703241] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 803.703431] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 803.703582] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 803.703858] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 803.704122] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.704289] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 803.704461] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 803.704629] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 803.704840] env[65107]: DEBUG nova.virt.hardware [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 803.705926] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505434a1-78e5-4642-8404-25d890db6a7f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.715731] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe312e7-37a0-42f5-ab2d-1820ae1cac32 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.845641] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102743, 'name': ReconfigVM_Task, 'duration_secs': 0.318353} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.846051] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.846369] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 803.850650] env[65107]: DEBUG nova.scheduler.client.report [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 803.892513] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.119531] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102744, 'name': Destroy_Task, 'duration_secs': 0.348329} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.119797] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroyed the VM [ 804.120049] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 804.120313] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-07397d1b-d5aa-4516-946b-9afe3f6606e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.128104] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 804.128104] env[65107]: value = "task-5102745" [ 804.128104] env[65107]: _type = "Task" [ 804.128104] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.136670] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102745, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.242537] env[65107]: DEBUG nova.network.neutron [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Successfully updated port: cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 804.356689] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.719s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.357208] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 804.360443] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f5aa99-8bec-43e7-b180-c02603b038ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.364952] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.523s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.365208] env[65107]: DEBUG nova.objects.instance [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lazy-loading 'resources' on Instance uuid 0e6c9f2d-4412-4d8d-8348-6552a34ab349 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.386819] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8e498f-fb5a-4a73-bb9e-5634b6447d63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.406317] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 804.588496] env[65107]: DEBUG nova.compute.manager [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Received event network-vif-plugged-cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 804.588829] env[65107]: DEBUG oslo_concurrency.lockutils [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.589097] env[65107]: DEBUG oslo_concurrency.lockutils [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.589453] env[65107]: DEBUG oslo_concurrency.lockutils [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.589593] env[65107]: DEBUG nova.compute.manager [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] No waiting events found dispatching network-vif-plugged-cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 804.589774] env[65107]: WARNING nova.compute.manager [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Received unexpected event network-vif-plugged-cef8867b-7dd6-425a-b5db-760494cf6eb4 for instance with vm_state building and task_state spawning. [ 804.589909] env[65107]: DEBUG nova.compute.manager [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Received event network-changed-cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 804.590080] env[65107]: DEBUG nova.compute.manager [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Refreshing instance network info cache due to event network-changed-cef8867b-7dd6-425a-b5db-760494cf6eb4. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 804.590271] env[65107]: DEBUG oslo_concurrency.lockutils [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.590450] env[65107]: DEBUG oslo_concurrency.lockutils [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.590573] env[65107]: DEBUG nova.network.neutron [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Refreshing network info cache for port cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 804.638922] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102745, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.745421] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.871489] env[65107]: DEBUG nova.compute.utils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 804.873187] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 804.873369] env[65107]: DEBUG nova.network.neutron [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 804.873884] env[65107]: WARNING neutronclient.v2_0.client [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.874494] env[65107]: WARNING neutronclient.v2_0.client [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.875159] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.875568] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.912706] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.913289] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.929249] env[65107]: DEBUG nova.policy [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c5b4a751c1e4a3086f7e3359042582b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b981c5f6b3b64476a2406cd9549f7d6b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 804.964106] env[65107]: DEBUG nova.network.neutron [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Port fcca9afe-edba-479c-93f6-1310627d946f binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 805.093748] env[65107]: WARNING neutronclient.v2_0.client [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.094798] env[65107]: WARNING openstack [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.094798] env[65107]: WARNING openstack [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.142398] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102745, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.164388] env[65107]: DEBUG nova.network.neutron [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 805.291515] env[65107]: DEBUG nova.network.neutron [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.331168] env[65107]: DEBUG nova.network.neutron [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Successfully created port: bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 805.384456] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 805.460755] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d64d89-4dba-4a1f-8ce9-5d401c946c8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.476255] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8428a8-0684-4520-a801-aaf1b9277d43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.514038] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769aa681-961e-4ff9-ba25-17f1903d1def {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.522802] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355c27d7-0733-4253-b4e2-e63fa5d28306 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.538377] env[65107]: DEBUG nova.compute.provider_tree [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.639921] env[65107]: DEBUG oslo_vmware.api [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102745, 'name': RemoveSnapshot_Task, 'duration_secs': 1.419218} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.640698] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 805.641058] env[65107]: INFO nova.compute.manager [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 13.92 seconds to snapshot the instance on the hypervisor. [ 805.795197] env[65107]: DEBUG oslo_concurrency.lockutils [req-d493f9b8-ef86-48a8-af75-35a3ac935ab3 req-0510bf9f-796b-4b00-9e2f-202cd688c61d service nova] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.795723] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.795948] env[65107]: DEBUG nova.network.neutron [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 805.803640] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.803869] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "8306ff90-b8bd-4270-8133-96abe483156b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.804105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "8306ff90-b8bd-4270-8133-96abe483156b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.804312] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "8306ff90-b8bd-4270-8133-96abe483156b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.804480] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "8306ff90-b8bd-4270-8133-96abe483156b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.806831] env[65107]: INFO nova.compute.manager [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Terminating instance [ 805.992064] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "369caf24-48a6-4640-8601-e30fd47453da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.992308] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.992512] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.041725] env[65107]: DEBUG nova.scheduler.client.report [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 806.201999] env[65107]: DEBUG nova.compute.manager [None req-1f692a04-1149-447f-bc29-bb072d031c8c tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Found 1 images (rotation: 2) {{(pid=65107) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 806.298731] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.299228] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.310535] env[65107]: DEBUG nova.compute.manager [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 806.310764] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.311735] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e74f40-2541-4905-90e6-255027b113e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.320910] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 806.321203] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49181f6a-5bf4-4987-bbd8-60c08341f8da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.328820] env[65107]: DEBUG oslo_vmware.api [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 806.328820] env[65107]: value = "task-5102746" [ 806.328820] env[65107]: _type = "Task" [ 806.328820] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.338539] env[65107]: DEBUG oslo_vmware.api [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.364311] env[65107]: DEBUG nova.network.neutron [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 806.395362] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 806.410254] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.410254] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.429213] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 806.429544] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 806.429646] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 806.430402] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 806.430977] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 806.431194] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 806.431456] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.431614] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 806.432066] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 806.432066] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 806.432200] env[65107]: DEBUG nova.virt.hardware [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 806.433076] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556ff1be-9d0f-49b9-9181-22d12c7121c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.445921] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e90ec2-cea4-4819-9e8c-366c298f8be1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.511790] env[65107]: WARNING neutronclient.v2_0.client [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.512631] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.513792] env[65107]: WARNING openstack [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.548523] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.183s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.553041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 40.534s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.553041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.553041] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 806.553041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.329s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.553553] env[65107]: INFO nova.compute.claims [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.557016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b982eb-f3f6-4dd7-b4e5-824222905500 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.566985] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a2d96a-d0d3-432b-88fc-16bdd9239aba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.589319] env[65107]: INFO nova.scheduler.client.report [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Deleted allocations for instance 0e6c9f2d-4412-4d8d-8348-6552a34ab349 [ 806.591105] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a966676-5435-4193-be4e-1dd1fae8d52a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.604141] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55600eeb-1210-4e87-9179-275d8cad04b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.642793] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177302MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 806.642793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.710373] env[65107]: DEBUG nova.network.neutron [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.785826] env[65107]: DEBUG nova.compute.manager [req-4d2eeb05-36dc-436d-bd2d-910973ddb54a req-69f90c12-202c-4256-b50b-383e13afc414 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Received event network-vif-plugged-bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 806.786050] env[65107]: DEBUG oslo_concurrency.lockutils [req-4d2eeb05-36dc-436d-bd2d-910973ddb54a req-69f90c12-202c-4256-b50b-383e13afc414 service nova] Acquiring lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.786263] env[65107]: DEBUG oslo_concurrency.lockutils [req-4d2eeb05-36dc-436d-bd2d-910973ddb54a req-69f90c12-202c-4256-b50b-383e13afc414 service nova] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.786431] env[65107]: DEBUG oslo_concurrency.lockutils [req-4d2eeb05-36dc-436d-bd2d-910973ddb54a req-69f90c12-202c-4256-b50b-383e13afc414 service nova] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.786598] env[65107]: DEBUG nova.compute.manager [req-4d2eeb05-36dc-436d-bd2d-910973ddb54a req-69f90c12-202c-4256-b50b-383e13afc414 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] No waiting events found dispatching network-vif-plugged-bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 806.786761] env[65107]: WARNING nova.compute.manager [req-4d2eeb05-36dc-436d-bd2d-910973ddb54a req-69f90c12-202c-4256-b50b-383e13afc414 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Received unexpected event network-vif-plugged-bec67e7b-de92-407d-aef5-04d4430d3ee9 for instance with vm_state building and task_state spawning. [ 806.839641] env[65107]: DEBUG oslo_vmware.api [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102746, 'name': PowerOffVM_Task, 'duration_secs': 0.274202} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.839986] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 806.840098] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 806.840354] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab5a9819-e6a2-45d9-9b6a-3c02b269e6e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.866235] env[65107]: DEBUG nova.network.neutron [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Successfully updated port: bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 806.873284] env[65107]: DEBUG nova.compute.manager [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 806.874255] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fef60b7-0c8b-4cd1-9bce-9498dc9c0dfc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.001169] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.055224] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 807.055224] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 807.055224] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Deleting the datastore file [datastore1] 8306ff90-b8bd-4270-8133-96abe483156b {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.055224] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbc5b308-3e32-4cef-886b-c725c55735bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.065599] env[65107]: DEBUG oslo_vmware.api [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for the task: (returnval){ [ 807.065599] env[65107]: value = "task-5102748" [ 807.065599] env[65107]: _type = "Task" [ 807.065599] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.071380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.071609] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.071823] env[65107]: DEBUG nova.network.neutron [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 807.078543] env[65107]: DEBUG oslo_vmware.api [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.100044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8dab05e-ba5e-4ca1-ad8c-a79516f21a5f tempest-ServersTestManualDisk-950056448 tempest-ServersTestManualDisk-950056448-project-member] Lock "0e6c9f2d-4412-4d8d-8348-6552a34ab349" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.280s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.213159] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.213534] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Instance network_info: |[{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 807.214019] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:3d:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f9ffee1-f413-4f28-8bc4-3fb2cf299789', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cef8867b-7dd6-425a-b5db-760494cf6eb4', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.221426] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating folder: Project (139a38d3b7bd4d3a8861d2e1e6c7a2c1). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.221662] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5cdc8258-5a06-4baa-9e36-730682ea2725 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.233513] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created folder: Project (139a38d3b7bd4d3a8861d2e1e6c7a2c1) in parent group-v992574. [ 807.233759] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating folder: Instances. Parent ref: group-v992721. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.234095] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9e52a15-7119-4f9d-ae80-337167a8bb14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.244773] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created folder: Instances in parent group-v992721. [ 807.245997] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 807.246271] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.246495] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11806188-8661-407f-a982-5cac702845da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.266865] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.266865] env[65107]: value = "task-5102751" [ 807.266865] env[65107]: _type = "Task" [ 807.266865] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.276011] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102751, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.369202] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.369202] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.369202] env[65107]: DEBUG nova.network.neutron [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 807.385324] env[65107]: INFO nova.compute.manager [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] instance snapshotting [ 807.386205] env[65107]: DEBUG nova.objects.instance [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'flavor' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.575254] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.576100] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.576548] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.588528] env[65107]: DEBUG oslo_vmware.api [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Task: {'id': task-5102748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149407} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.588774] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 807.588958] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 807.589150] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.589596] env[65107]: INFO nova.compute.manager [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Took 1.28 seconds to destroy the instance on the hypervisor. [ 807.589596] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 807.589757] env[65107]: DEBUG nova.compute.manager [-] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 807.589856] env[65107]: DEBUG nova.network.neutron [-] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 807.590116] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.590629] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.590884] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.716737] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.782539] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102751, 'name': CreateVM_Task, 'duration_secs': 0.440709} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.782823] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.784207] env[65107]: WARNING neutronclient.v2_0.client [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.784207] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.784207] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.784207] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 807.784692] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16b90a40-d913-4894-8b8d-ef5283594c94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.790741] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 807.790741] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e38df-e627-d271-c293-2bfc8c8d6522" [ 807.790741] env[65107]: _type = "Task" [ 807.790741] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.804194] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e38df-e627-d271-c293-2bfc8c8d6522, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.871917] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.872345] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.892674] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e40bc8a-0c4f-48d9-b8ee-e70da93ca9f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.920516] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbb1903-1953-4ad3-b386-834c7df8ce56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.150056] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.150588] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.159087] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1fcdd3-546b-4571-86ec-133976b480e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.167210] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56821951-f5ed-4efa-bd13-129dba652ff1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.202118] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb172ec-d61b-4982-bed5-183eb770526c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.210414] env[65107]: DEBUG nova.network.neutron [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 808.213775] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46b476f-3bd2-42c0-8a93-e7bec9891d38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.229115] env[65107]: DEBUG nova.compute.provider_tree [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.306341] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528e38df-e627-d271-c293-2bfc8c8d6522, 'name': SearchDatastore_Task, 'duration_secs': 0.015674} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.307331] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.307678] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.315211] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.315963] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.315963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.315963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.316346] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.317073] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07f7b607-5fb4-47a6-812e-15fa0e78001f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.327851] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.328100] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.328998] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f1bd7f2-cfc1-4975-83c2-ec6428fa8334 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.336384] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 808.336384] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524e3699-db09-0707-03b0-f2709ddd52df" [ 808.336384] env[65107]: _type = "Task" [ 808.336384] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.346029] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524e3699-db09-0707-03b0-f2709ddd52df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.411879] env[65107]: WARNING neutronclient.v2_0.client [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.411879] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.411879] env[65107]: WARNING openstack [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.434385] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 808.434385] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-45ade310-99c1-4dbb-a94a-812180b19097 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.442246] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 808.442246] env[65107]: value = "task-5102752" [ 808.442246] env[65107]: _type = "Task" [ 808.442246] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.452465] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102752, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.542804] env[65107]: WARNING neutronclient.v2_0.client [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.543595] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.544012] env[65107]: WARNING openstack [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.598574] env[65107]: DEBUG nova.network.neutron [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 808.702201] env[65107]: DEBUG nova.network.neutron [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updating instance_info_cache with network_info: [{"id": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "address": "fa:16:3e:0f:eb:48", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbec67e7b-de", "ovs_interfaceid": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 808.732689] env[65107]: DEBUG nova.scheduler.client.report [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 808.764469] env[65107]: DEBUG nova.network.neutron [-] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 808.854914] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524e3699-db09-0707-03b0-f2709ddd52df, 'name': SearchDatastore_Task, 'duration_secs': 0.015202} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.854914] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f11e6f4b-abc7-4542-a74c-09a7c30f8b75 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.863801] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 808.863801] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5279419a-4b0a-7dad-1904-18cf7c6d9a6e" [ 808.863801] env[65107]: _type = "Task" [ 808.863801] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.882237] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5279419a-4b0a-7dad-1904-18cf7c6d9a6e, 'name': SearchDatastore_Task, 'duration_secs': 0.012877} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.882606] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.883046] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a/cd481deb-b12c-47ab-9ab1-0c4bdbb2024a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.883046] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f220b13-e5fc-4788-b09d-7d1634984476 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.895460] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 808.895460] env[65107]: value = "task-5102753" [ 808.895460] env[65107]: _type = "Task" [ 808.895460] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.905176] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.963297] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102752, 'name': CreateSnapshot_Task, 'duration_secs': 0.475539} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.963703] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 808.964854] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d88e07-8983-4576-96c2-a5e1fcf13bd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.102037] env[65107]: DEBUG oslo_concurrency.lockutils [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.114986] env[65107]: DEBUG nova.compute.manager [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Received event network-changed-bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 809.115297] env[65107]: DEBUG nova.compute.manager [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Refreshing instance network info cache due to event network-changed-bec67e7b-de92-407d-aef5-04d4430d3ee9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 809.115528] env[65107]: DEBUG oslo_concurrency.lockutils [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Acquiring lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.208568] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.209068] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Instance network_info: |[{"id": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "address": "fa:16:3e:0f:eb:48", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbec67e7b-de", "ovs_interfaceid": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 809.209509] env[65107]: DEBUG oslo_concurrency.lockutils [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Acquired lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.209623] env[65107]: DEBUG nova.network.neutron [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Refreshing network info cache for port bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 809.211542] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:eb:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4020f51-6e46-4b73-a79e-9fe3fd51b917', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bec67e7b-de92-407d-aef5-04d4430d3ee9', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.219730] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Creating folder: Project (b981c5f6b3b64476a2406cd9549f7d6b). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.222055] env[65107]: WARNING neutronclient.v2_0.client [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.222164] env[65107]: WARNING openstack [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.222902] env[65107]: WARNING openstack [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.230983] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e70cf7ae-7f29-4444-ba41-985e33deed68 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.240962] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.689s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.241694] env[65107]: DEBUG nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 809.246301] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.100s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.246551] env[65107]: DEBUG nova.objects.instance [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lazy-loading 'resources' on Instance uuid c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.248569] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Created folder: Project (b981c5f6b3b64476a2406cd9549f7d6b) in parent group-v992574. [ 809.248761] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Creating folder: Instances. Parent ref: group-v992725. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.249407] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aced723d-f1b2-40fa-8d46-c5bbd77dc380 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.264417] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Created folder: Instances in parent group-v992725. [ 809.264698] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 809.264923] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.265182] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4771b49f-d37a-418e-92d8-fd3d262a8792 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.280842] env[65107]: INFO nova.compute.manager [-] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Took 1.69 seconds to deallocate network for instance. [ 809.293437] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.293437] env[65107]: value = "task-5102756" [ 809.293437] env[65107]: _type = "Task" [ 809.293437] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.305080] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102756, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.385055] env[65107]: WARNING openstack [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.385502] env[65107]: WARNING openstack [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.407449] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102753, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.457764] env[65107]: WARNING neutronclient.v2_0.client [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.458437] env[65107]: WARNING openstack [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.458779] env[65107]: WARNING openstack [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.488438] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 809.488776] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e483d02b-128c-4b27-964f-6f3464ea7a52 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.501022] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 809.501022] env[65107]: value = "task-5102757" [ 809.501022] env[65107]: _type = "Task" [ 809.501022] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.508108] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102757, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.559691] env[65107]: DEBUG nova.network.neutron [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updated VIF entry in instance network info cache for port bec67e7b-de92-407d-aef5-04d4430d3ee9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 809.560686] env[65107]: DEBUG nova.network.neutron [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updating instance_info_cache with network_info: [{"id": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "address": "fa:16:3e:0f:eb:48", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbec67e7b-de", "ovs_interfaceid": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.628463] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9b47ac-52e9-4392-9d04-ddadca3d4e83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.653177] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa91cb5d-3e08-4f94-b862-7a3b98e03172 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.662107] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 809.754355] env[65107]: DEBUG nova.compute.utils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 809.756304] env[65107]: DEBUG nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 809.791188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.806572] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102756, 'name': CreateVM_Task, 'duration_secs': 0.411799} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.809796] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.810504] env[65107]: WARNING neutronclient.v2_0.client [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.810863] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.811025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.811338] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 809.811596] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4efd0017-066d-4f5e-9614-fdd1cd6d26e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.817481] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 809.817481] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241d717-504d-0345-7e91-3204fff92502" [ 809.817481] env[65107]: _type = "Task" [ 809.817481] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.827840] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241d717-504d-0345-7e91-3204fff92502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.910947] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533049} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.911396] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a/cd481deb-b12c-47ab-9ab1-0c4bdbb2024a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.911507] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.911943] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-665d5e44-6ce8-4c8e-b087-22a1edf489ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.920623] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 809.920623] env[65107]: value = "task-5102758" [ 809.920623] env[65107]: _type = "Task" [ 809.920623] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.935765] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102758, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.015226] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102757, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.063644] env[65107]: DEBUG oslo_concurrency.lockutils [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] Releasing lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.063733] env[65107]: DEBUG nova.compute.manager [req-7e32b2ef-bde2-45b9-bc50-3d46746e92a3 req-c9b9ddc7-a054-4f91-b1f1-6fdaeeab71b5 service nova] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Received event network-vif-deleted-e41be772-a45d-410b-acaf-09d30f9afa9d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 810.168781] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.169362] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5715a4e6-58e8-4f4e-ae63-c2c9507dbb89 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.176854] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 810.176854] env[65107]: value = "task-5102759" [ 810.176854] env[65107]: _type = "Task" [ 810.176854] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.192206] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102759, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.258360] env[65107]: DEBUG nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 810.298113] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d41dad-21c5-4d7e-b586-361b1d6d1967 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.306329] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dead6978-a6d2-4677-86e9-cd0e75aed0ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.347131] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13da946-5d4c-462a-adaa-6a72f9dd1e11 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.356446] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5241d717-504d-0345-7e91-3204fff92502, 'name': SearchDatastore_Task, 'duration_secs': 0.010684} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.358768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.359059] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.359309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.359454] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.359629] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.360391] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-110faade-cfc8-48c3-a18a-ae652345454e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.363252] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3772892f-739b-4063-8d89-4076a0eac10d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.378634] env[65107]: DEBUG nova.compute.provider_tree [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.384020] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.384020] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.384020] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7cd0161-071c-4c8c-9c63-c2f7bb73171b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.387839] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 810.387839] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526d07b5-8735-f09b-7ac9-ea07567167af" [ 810.387839] env[65107]: _type = "Task" [ 810.387839] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.397365] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526d07b5-8735-f09b-7ac9-ea07567167af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.431899] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102758, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081004} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.432352] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.433330] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c7d7e2-c9ee-4973-bb7a-92abafab0d92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.459033] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a/cd481deb-b12c-47ab-9ab1-0c4bdbb2024a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.459724] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0725f041-c7b5-447e-b55a-2a8bdc46ab3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.482106] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 810.482106] env[65107]: value = "task-5102760" [ 810.482106] env[65107]: _type = "Task" [ 810.482106] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.491462] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102760, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.509811] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102757, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.689935] env[65107]: DEBUG oslo_vmware.api [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102759, 'name': PowerOnVM_Task, 'duration_secs': 0.420568} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.690958] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.690958] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-38b3c3a9-be92-408a-872c-660a4cd9db7c tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance '369caf24-48a6-4640-8601-e30fd47453da' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 810.883931] env[65107]: DEBUG nova.scheduler.client.report [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 810.905449] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526d07b5-8735-f09b-7ac9-ea07567167af, 'name': SearchDatastore_Task, 'duration_secs': 0.010244} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.910346] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57768a4e-2e8a-43d6-9e46-93e3a4e46872 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.918339] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 810.918339] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a579c4-228a-67e9-d7a5-81f157c57a15" [ 810.918339] env[65107]: _type = "Task" [ 810.918339] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.927537] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a579c4-228a-67e9-d7a5-81f157c57a15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.993116] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102760, 'name': ReconfigVM_Task, 'duration_secs': 0.411168} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.993116] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Reconfigured VM instance instance-00000031 to attach disk [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a/cd481deb-b12c-47ab-9ab1-0c4bdbb2024a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.993827] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff5b16fb-4f1a-4ee3-8d54-9ff85de35080 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.000873] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 811.000873] env[65107]: value = "task-5102761" [ 811.000873] env[65107]: _type = "Task" [ 811.000873] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.013249] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102757, 'name': CloneVM_Task, 'duration_secs': 1.336241} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.016752] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created linked-clone VM from snapshot [ 811.017158] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102761, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.017978] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030984dc-bbeb-4c41-bced-a71dea3ae61a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.026634] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Uploading image 441569b7-9bcf-462e-b53e-b76f2b2cd9a4 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 811.054399] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 811.054399] env[65107]: value = "vm-992728" [ 811.054399] env[65107]: _type = "VirtualMachine" [ 811.054399] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 811.054703] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-89ca361c-6836-4d42-b559-101028569542 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.063280] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease: (returnval){ [ 811.063280] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010cff-6445-c275-85f8-a028a89d8683" [ 811.063280] env[65107]: _type = "HttpNfcLease" [ 811.063280] env[65107]: } obtained for exporting VM: (result){ [ 811.063280] env[65107]: value = "vm-992728" [ 811.063280] env[65107]: _type = "VirtualMachine" [ 811.063280] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 811.063643] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the lease: (returnval){ [ 811.063643] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010cff-6445-c275-85f8-a028a89d8683" [ 811.063643] env[65107]: _type = "HttpNfcLease" [ 811.063643] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 811.071887] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 811.071887] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010cff-6445-c275-85f8-a028a89d8683" [ 811.071887] env[65107]: _type = "HttpNfcLease" [ 811.071887] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 811.270350] env[65107]: DEBUG nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 811.296690] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 811.296973] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 811.297146] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 811.297333] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 811.297831] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 811.298066] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 811.298307] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.298469] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 811.298635] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 811.298809] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 811.298990] env[65107]: DEBUG nova.virt.hardware [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 811.300012] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76b9c0d-a4cd-4bde-bf1a-b305d10cf5e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.308858] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1f5ec5-26e2-4e11-af74-4b10bbee5450 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.324684] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.332306] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Creating folder: Project (f2746b61170e4305836b12d641fec06e). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.332715] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03d89624-0eea-4ab6-8cd0-f1df5bea230f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.344816] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Created folder: Project (f2746b61170e4305836b12d641fec06e) in parent group-v992574. [ 811.345081] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Creating folder: Instances. Parent ref: group-v992729. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.345336] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68a2fe55-dacc-464c-a3df-6d399271a417 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.359894] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Created folder: Instances in parent group-v992729. [ 811.360196] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 811.360426] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.360653] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-553b4ecd-7536-4f7d-ab0c-48b0d103286c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.379801] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.379801] env[65107]: value = "task-5102765" [ 811.379801] env[65107]: _type = "Task" [ 811.379801] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.388415] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102765, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.390432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.144s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 811.392912] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.929s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 811.393710] env[65107]: DEBUG nova.objects.instance [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lazy-loading 'resources' on Instance uuid 90706f49-af87-4038-ae76-02a95e4a12e6 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.425251] env[65107]: INFO nova.scheduler.client.report [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleted allocations for instance c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0 [ 811.432270] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a579c4-228a-67e9-d7a5-81f157c57a15, 'name': SearchDatastore_Task, 'duration_secs': 0.037266} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.438035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.438035] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 179e613b-e5a2-4fbc-8fa6-b72769425ff3/179e613b-e5a2-4fbc-8fa6-b72769425ff3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.438035] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2dc532fa-3db5-427e-a455-c8e0f0c641aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.444394] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 811.444394] env[65107]: value = "task-5102766" [ 811.444394] env[65107]: _type = "Task" [ 811.444394] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.458862] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.516472] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102761, 'name': Rename_Task, 'duration_secs': 0.161806} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.516472] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.516472] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e020ff5-099d-43a1-b138-1a47f4c73833 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.526029] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 811.526029] env[65107]: value = "task-5102767" [ 811.526029] env[65107]: _type = "Task" [ 811.526029] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.538959] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.581024] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 811.581024] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010cff-6445-c275-85f8-a028a89d8683" [ 811.581024] env[65107]: _type = "HttpNfcLease" [ 811.581024] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 811.581384] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 811.581384] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010cff-6445-c275-85f8-a028a89d8683" [ 811.581384] env[65107]: _type = "HttpNfcLease" [ 811.581384] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 811.582351] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b040b063-5dbf-4807-998e-97c0c7c876a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.591812] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e82c88-9651-d913-8977-9d9b975d7788/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 811.592059] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e82c88-9651-d913-8977-9d9b975d7788/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 811.706717] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0332ff86-0d0b-4383-874d-cf656f6a6129 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.898169] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102765, 'name': CreateVM_Task, 'duration_secs': 0.314195} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.898840] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.902631] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.902825] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.905823] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 811.905823] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b358523c-8be8-446b-9bf4-fa2cc36281fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.913095] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 811.913095] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fd9f38-614f-f550-d96d-ea0b6e59aeeb" [ 811.913095] env[65107]: _type = "Task" [ 811.913095] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.926706] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fd9f38-614f-f550-d96d-ea0b6e59aeeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.941900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e8d00c62-624c-4663-b023-8b3e2446266a tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.327s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 811.961784] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102766, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.040124] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102767, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.427148] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fd9f38-614f-f550-d96d-ea0b6e59aeeb, 'name': SearchDatastore_Task, 'duration_secs': 0.06852} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.427237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.427619] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.427901] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.428087] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 812.428301] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.428768] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc682d7b-0317-4f96-80e9-3ba4a0baad09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.442403] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.442714] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.443637] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8f80d85-4996-4820-84bd-a723ecfc0000 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.453721] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 812.453721] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529524d4-8fae-4746-3f58-49d6bc7c7719" [ 812.453721] env[65107]: _type = "Task" [ 812.453721] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.464785] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628785} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.465658] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 179e613b-e5a2-4fbc-8fa6-b72769425ff3/179e613b-e5a2-4fbc-8fa6-b72769425ff3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.465731] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.466315] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-975ec96b-cde5-4b34-8fc5-11d73ab5500f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.471508] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529524d4-8fae-4746-3f58-49d6bc7c7719, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.477939] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 812.477939] env[65107]: value = "task-5102768" [ 812.477939] env[65107]: _type = "Task" [ 812.477939] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.488978] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.498181] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-d2d0e18c-4839-416c-b363-b93996e8ba7f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.498181] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-d2d0e18c-4839-416c-b363-b93996e8ba7f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.499282] env[65107]: DEBUG nova.objects.instance [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'flavor' on Instance uuid d2d0e18c-4839-416c-b363-b93996e8ba7f {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.507528] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2860ad6c-9768-4d9a-9756-c63debd41711 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.518688] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b774b0cf-bb19-4f5f-ac14-cf5ebc98abb5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.558913] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4321f849-590f-447e-8864-a219f97ed7d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.570430] env[65107]: DEBUG oslo_vmware.api [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5102767, 'name': PowerOnVM_Task, 'duration_secs': 0.628044} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.571985] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910dd5fd-d6d4-40ae-b7f9-475c9f2489ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.576700] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.577170] env[65107]: INFO nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Took 8.90 seconds to spawn the instance on the hypervisor. [ 812.577551] env[65107]: DEBUG nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 812.578533] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc26247-76cf-4eaa-a981-0b4d5a437c56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.594645] env[65107]: DEBUG nova.compute.provider_tree [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.752187] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.806370] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.806751] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.807032] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.852399] env[65107]: DEBUG nova.network.neutron [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Port fcca9afe-edba-479c-93f6-1310627d946f binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 812.852815] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.852993] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 812.853128] env[65107]: DEBUG nova.network.neutron [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 812.970521] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529524d4-8fae-4746-3f58-49d6bc7c7719, 'name': SearchDatastore_Task, 'duration_secs': 0.030783} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.972581] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-840e045b-93ae-4c10-ac59-8ab1fac957b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.980529] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 812.980529] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5283ad87-4490-4ec3-f815-6c2c05120455" [ 812.980529] env[65107]: _type = "Task" [ 812.980529] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.000550] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5283ad87-4490-4ec3-f815-6c2c05120455, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.000924] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.225607} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.001372] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.002284] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd798a5-2a51-4911-8838-0c1404a2d832 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.006941] env[65107]: WARNING neutronclient.v2_0.client [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.007743] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.008286] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.037433] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 179e613b-e5a2-4fbc-8fa6-b72769425ff3/179e613b-e5a2-4fbc-8fa6-b72769425ff3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.037935] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-244884e0-0afa-45bb-9931-895da66f9aa9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.062488] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 813.062488] env[65107]: value = "task-5102769" [ 813.062488] env[65107]: _type = "Task" [ 813.062488] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.073504] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102769, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.103702] env[65107]: DEBUG nova.scheduler.client.report [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 813.119529] env[65107]: INFO nova.compute.manager [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Took 56.26 seconds to build instance. [ 813.132707] env[65107]: DEBUG nova.objects.instance [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'pci_requests' on Instance uuid d2d0e18c-4839-416c-b363-b93996e8ba7f {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.356486] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.359076] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.361440] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.524490] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5283ad87-4490-4ec3-f815-6c2c05120455, 'name': SearchDatastore_Task, 'duration_secs': 0.016204} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.524490] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.524490] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.524490] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9a021ec-84be-46fb-9de4-63417e6675bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.531955] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 813.531955] env[65107]: value = "task-5102770" [ 813.531955] env[65107]: _type = "Task" [ 813.531955] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.542084] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.543029] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.554050] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.574035] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102769, 'name': ReconfigVM_Task, 'duration_secs': 0.437608} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.574342] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 179e613b-e5a2-4fbc-8fa6-b72769425ff3/179e613b-e5a2-4fbc-8fa6-b72769425ff3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.575107] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d025389-d286-4ce4-a621-d0bed5f55148 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.589032] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 813.589032] env[65107]: value = "task-5102771" [ 813.589032] env[65107]: _type = "Task" [ 813.589032] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.603227] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102771, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.612038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.219s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.616871] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.617735] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.618469] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.628218] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.060s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.630166] env[65107]: INFO nova.compute.claims [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.633514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7f2f64e2-0dd7-4e9f-905f-06f0cd3dc633 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.789s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.634714] env[65107]: DEBUG nova.objects.base [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 813.634964] env[65107]: DEBUG nova.network.neutron [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 813.635375] env[65107]: WARNING neutronclient.v2_0.client [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.635821] env[65107]: WARNING neutronclient.v2_0.client [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.636482] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.636899] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.657948] env[65107]: INFO nova.scheduler.client.report [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Deleted allocations for instance 90706f49-af87-4038-ae76-02a95e4a12e6 [ 813.721758] env[65107]: DEBUG nova.policy [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 813.757881] env[65107]: DEBUG nova.network.neutron [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 814.005035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.005035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.051084] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102770, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.109177] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102771, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.164989] env[65107]: DEBUG nova.network.neutron [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Successfully created port: 97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 814.171680] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a203ca08-0481-41af-918a-731eaabf1135 tempest-ListServersNegativeTestJSON-1289631193 tempest-ListServersNegativeTestJSON-1289631193-project-member] Lock "90706f49-af87-4038-ae76-02a95e4a12e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.462s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.261230] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.509714] env[65107]: DEBUG nova.compute.utils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 814.545847] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102770, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628834} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.547678] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.547678] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.550365] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a917337e-3baa-4efb-958d-704fe3db7674 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.553353] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "88cac920-3c1e-41dc-9e52-f51f48a80f56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.553651] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.562765] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 814.562765] env[65107]: value = "task-5102772" [ 814.562765] env[65107]: _type = "Task" [ 814.562765] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.574918] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.602678] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102771, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.764188] env[65107]: DEBUG nova.compute.manager [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65107) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 814.764437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.012657] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.056535] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 815.074247] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074807} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.074536] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.075368] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc112c2d-b23a-4fb2-8646-51634662eb61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.099439] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.103926] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c58079a7-e325-4886-bcd5-c5e89824a4be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.131878] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102771, 'name': Rename_Task, 'duration_secs': 1.232283} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.140452] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.140452] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 815.140452] env[65107]: value = "task-5102773" [ 815.140452] env[65107]: _type = "Task" [ 815.140452] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.140452] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7756954-b1c1-4bde-b819-79f708a22c49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.156516] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 815.156516] env[65107]: value = "task-5102774" [ 815.156516] env[65107]: _type = "Task" [ 815.156516] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.156989] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102773, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.166299] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.171718] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3158cbc5-1a67-488d-912d-2ab3db68500f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.182247] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90edd07d-bdae-4d00-9f87-c803b3df4c4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.215772] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b093d5-f77c-4eb4-8560-38cd5c567974 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.224390] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c7b649-885e-45a4-9985-68d05d7bcc37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.242352] env[65107]: DEBUG nova.compute.provider_tree [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.577700] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.652311] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102773, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.666993] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102774, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.699734] env[65107]: DEBUG nova.network.neutron [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Successfully updated port: 97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 815.745716] env[65107]: DEBUG nova.scheduler.client.report [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 816.092896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.092896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.093309] env[65107]: INFO nova.compute.manager [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Attaching volume aae9fa95-1cf2-433e-94d4-b3c0b1119742 to /dev/sdb [ 816.132982] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb9c0e8-dcd0-449c-aa8a-6ea953133dc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.141451] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ba9ff9-db4c-4cc9-9c8a-6d11173c26e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.159576] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102773, 'name': ReconfigVM_Task, 'duration_secs': 0.736417} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.167430] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.168279] env[65107]: DEBUG nova.virt.block_device [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updating existing volume attachment record: 190bdb30-76c1-4389-bad8-70c14a6c3c55 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 816.171041] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd86ef99-36e1-4dae-b2bf-75273c7a86f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.180656] env[65107]: DEBUG oslo_vmware.api [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102774, 'name': PowerOnVM_Task, 'duration_secs': 0.832053} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.182207] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.182524] env[65107]: INFO nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Took 9.79 seconds to spawn the instance on the hypervisor. [ 816.182736] env[65107]: DEBUG nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 816.183170] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 816.183170] env[65107]: value = "task-5102775" [ 816.183170] env[65107]: _type = "Task" [ 816.183170] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.183913] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b1ea20-d1cc-4122-afb7-631b128096e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.201698] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102775, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.202357] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.202642] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.202642] env[65107]: DEBUG nova.network.neutron [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 816.254022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.254022] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 816.255781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.782s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.256302] env[65107]: DEBUG nova.objects.instance [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lazy-loading 'resources' on Instance uuid 6563cbda-5fd3-4640-9e9a-95904f3f8335 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.701893] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102775, 'name': Rename_Task, 'duration_secs': 0.296725} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.706477] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.710905] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.711342] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.719241] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae626341-4eb5-429d-b695-6ce694359fa1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.721388] env[65107]: INFO nova.compute.manager [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Took 57.88 seconds to build instance. [ 816.728682] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 816.728682] env[65107]: value = "task-5102779" [ 816.728682] env[65107]: _type = "Task" [ 816.728682] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.739661] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102779, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.760370] env[65107]: DEBUG nova.compute.utils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 816.762858] env[65107]: WARNING nova.network.neutron [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] 40a76aa5-ac46-4067-a98d-cd60410bf479 already exists in list: networks containing: ['40a76aa5-ac46-4067-a98d-cd60410bf479']. ignoring it [ 816.767855] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 816.768221] env[65107]: DEBUG nova.network.neutron [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 816.768506] env[65107]: WARNING neutronclient.v2_0.client [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.768811] env[65107]: WARNING neutronclient.v2_0.client [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.769415] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.769939] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.815587] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.815970] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.826110] env[65107]: DEBUG nova.policy [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad5a8049a2cf40cc8e710a78ea210ff5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccad355f4d8546d2863753bfb1c86ba9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 816.865452] env[65107]: DEBUG nova.compute.manager [req-5bd754c8-50c9-47d2-b502-afc53ceed6ab req-d2dfd865-9a75-4923-ad80-0694e04607ce service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-vif-plugged-97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 816.865452] env[65107]: DEBUG oslo_concurrency.lockutils [req-5bd754c8-50c9-47d2-b502-afc53ceed6ab req-d2dfd865-9a75-4923-ad80-0694e04607ce service nova] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.865452] env[65107]: DEBUG oslo_concurrency.lockutils [req-5bd754c8-50c9-47d2-b502-afc53ceed6ab req-d2dfd865-9a75-4923-ad80-0694e04607ce service nova] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.865452] env[65107]: DEBUG oslo_concurrency.lockutils [req-5bd754c8-50c9-47d2-b502-afc53ceed6ab req-d2dfd865-9a75-4923-ad80-0694e04607ce service nova] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.865452] env[65107]: DEBUG nova.compute.manager [req-5bd754c8-50c9-47d2-b502-afc53ceed6ab req-d2dfd865-9a75-4923-ad80-0694e04607ce service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] No waiting events found dispatching network-vif-plugged-97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 816.865452] env[65107]: WARNING nova.compute.manager [req-5bd754c8-50c9-47d2-b502-afc53ceed6ab req-d2dfd865-9a75-4923-ad80-0694e04607ce service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received unexpected event network-vif-plugged-97579a1b-69c4-4b0c-b832-426f4d60592d for instance with vm_state active and task_state None. [ 816.887350] env[65107]: WARNING neutronclient.v2_0.client [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.888576] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.888576] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.916866] env[65107]: DEBUG nova.compute.manager [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Received event network-changed-cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 816.917086] env[65107]: DEBUG nova.compute.manager [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Refreshing instance network info cache due to event network-changed-cef8867b-7dd6-425a-b5db-760494cf6eb4. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 816.917303] env[65107]: DEBUG oslo_concurrency.lockutils [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.917446] env[65107]: DEBUG oslo_concurrency.lockutils [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.917665] env[65107]: DEBUG nova.network.neutron [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Refreshing network info cache for port cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 817.195880] env[65107]: DEBUG nova.network.neutron [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Successfully created port: 487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 817.223919] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a994ddea-6ea7-42d5-ac4d-c950a5a5ef54 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.406s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.248444] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102779, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.279303] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 817.361283] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1f3d27-32e5-4bae-a538-341dac6a7343 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.371524] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed853476-c2fd-4b3f-b563-73068f9ed9d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.411465] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86a91b2-613a-4830-b88a-bd912f1b8bce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.420768] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07240681-8471-4e5f-b424-cf218326bbd4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.425496] env[65107]: WARNING neutronclient.v2_0.client [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.426193] env[65107]: WARNING openstack [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.426513] env[65107]: WARNING openstack [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.446035] env[65107]: DEBUG nova.compute.provider_tree [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.743140] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102779, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.950918] env[65107]: DEBUG nova.scheduler.client.report [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.245559] env[65107]: DEBUG oslo_vmware.api [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102779, 'name': PowerOnVM_Task, 'duration_secs': 1.207103} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.246016] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.246505] env[65107]: INFO nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Took 6.98 seconds to spawn the instance on the hypervisor. [ 818.246625] env[65107]: DEBUG nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 818.247846] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8aecbe-0bbf-431a-a623-5844151da35c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.289924] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 818.301465] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.301465] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.323520] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 818.323736] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 818.323936] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 818.324067] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 818.324231] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 818.324501] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 818.324599] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.324756] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 818.324919] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 818.325128] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 818.325284] env[65107]: DEBUG nova.virt.hardware [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 818.327268] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7d8ad1-9ef2-4c8b-9aab-9ec19d1a6ee2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.338293] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1141e80-5fd4-4b47-bb5f-fcb175e89569 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.464488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.205s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.464488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.762s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 818.466217] env[65107]: DEBUG nova.objects.instance [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lazy-loading 'resources' on Instance uuid f58c3088-c821-4eec-be0a-81221debc98e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.492955] env[65107]: INFO nova.scheduler.client.report [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Deleted allocations for instance 6563cbda-5fd3-4640-9e9a-95904f3f8335 [ 818.623779] env[65107]: WARNING openstack [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.624291] env[65107]: WARNING openstack [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.766827] env[65107]: INFO nova.compute.manager [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Took 49.57 seconds to build instance. [ 818.928492] env[65107]: DEBUG nova.network.neutron [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Successfully updated port: 487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 818.972336] env[65107]: WARNING neutronclient.v2_0.client [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.974203] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.974626] env[65107]: WARNING openstack [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.014514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-16aff422-d46e-46d6-aed2-81e3ec8515b8 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "6563cbda-5fd3-4640-9e9a-95904f3f8335" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.564s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 819.068563] env[65107]: WARNING neutronclient.v2_0.client [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.069392] env[65107]: WARNING openstack [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.069715] env[65107]: WARNING openstack [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.142346] env[65107]: DEBUG nova.network.neutron [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "97579a1b-69c4-4b0c-b832-426f4d60592d", "address": "fa:16:3e:c4:9c:4c", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97579a1b-69", "ovs_interfaceid": "97579a1b-69c4-4b0c-b832-426f4d60592d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.267871] env[65107]: DEBUG nova.network.neutron [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updated VIF entry in instance network info cache for port cef8867b-7dd6-425a-b5db-760494cf6eb4. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 819.268349] env[65107]: DEBUG nova.network.neutron [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.269999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b4ad8f2d-3427-448c-a4e3-5e6a918a0810 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "780f910f-4f73-41ba-a795-0daae3097314" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.076s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 819.432643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "refresh_cache-5fccc802-f5ac-4450-8246-4cf9a5371046" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.432875] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquired lock "refresh_cache-5fccc802-f5ac-4450-8246-4cf9a5371046" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.433144] env[65107]: DEBUG nova.network.neutron [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 819.567146] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f308bdab-f993-44b4-80ef-63800040f487 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.576555] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32e83a5-d0cb-4f7a-aec9-a6eda2a8ebc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.611893] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37a1d4b-3c3b-433f-9e91-59890540a5b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.621180] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23adf5b-3b6f-4669-a50e-a333f7f5094d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.640723] env[65107]: DEBUG nova.compute.provider_tree [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.649926] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 819.650389] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.650614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.654106] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e7abd4-ebd3-40e6-82f2-c5d54ec4584d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.680422] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 819.680422] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 819.680547] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 819.680856] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 819.680980] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 819.681319] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 819.681585] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.681817] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 819.682099] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 819.682376] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 819.682658] env[65107]: DEBUG nova.virt.hardware [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 819.694090] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Reconfiguring VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 819.695154] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41300ac7-4e9f-412a-bff9-1e1001a12d89 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.726125] env[65107]: DEBUG oslo_vmware.api [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 819.726125] env[65107]: value = "task-5102781" [ 819.726125] env[65107]: _type = "Task" [ 819.726125] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.741285] env[65107]: DEBUG oslo_vmware.api [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102781, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.772219] env[65107]: DEBUG oslo_concurrency.lockutils [req-8fc80cce-57ab-42ef-9820-3a4e0ba4d2b2 req-c15e57e9-8f44-4930-998e-baf092bf1b26 service nova] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 819.936017] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.936531] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.984927] env[65107]: DEBUG nova.network.neutron [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 820.021030] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.022796] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.128177] env[65107]: WARNING neutronclient.v2_0.client [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 820.128881] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.129256] env[65107]: WARNING openstack [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.144145] env[65107]: DEBUG nova.scheduler.client.report [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 820.187870] env[65107]: DEBUG nova.compute.manager [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-changed-97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 820.188166] env[65107]: DEBUG nova.compute.manager [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Refreshing instance network info cache due to event network-changed-97579a1b-69c4-4b0c-b832-426f4d60592d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 820.188370] env[65107]: DEBUG oslo_concurrency.lockutils [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Acquiring lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.188506] env[65107]: DEBUG oslo_concurrency.lockutils [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Acquired lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 820.188717] env[65107]: DEBUG nova.network.neutron [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Refreshing network info cache for port 97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 820.244694] env[65107]: DEBUG oslo_vmware.api [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102781, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.249295] env[65107]: DEBUG nova.network.neutron [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Updating instance_info_cache with network_info: [{"id": "487dbf21-efa0-47ad-8919-f245c183a8ab", "address": "fa:16:3e:d1:d0:a6", "network": {"id": "1a32c316-9a81-4aa2-a49d-80ff12af4579", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1064776770-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccad355f4d8546d2863753bfb1c86ba9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487dbf21-ef", "ovs_interfaceid": "487dbf21-efa0-47ad-8919-f245c183a8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 820.651371] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.187s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.654658] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.333s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.655176] env[65107]: DEBUG nova.objects.instance [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lazy-loading 'resources' on Instance uuid 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.688716] env[65107]: INFO nova.scheduler.client.report [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted allocations for instance f58c3088-c821-4eec-be0a-81221debc98e [ 820.698498] env[65107]: WARNING neutronclient.v2_0.client [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 820.702279] env[65107]: WARNING openstack [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.702279] env[65107]: WARNING openstack [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.723738] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 820.724171] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992733', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'name': 'volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '82e8c093-b718-4d38-9682-ba2e710e4b93', 'attached_at': '', 'detached_at': '', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'serial': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 820.725573] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d1788b-5022-4636-a9ca-b33c86daa0ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.751717] env[65107]: DEBUG oslo_vmware.api [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102781, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.752242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Releasing lock "refresh_cache-5fccc802-f5ac-4450-8246-4cf9a5371046" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.753093] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Instance network_info: |[{"id": "487dbf21-efa0-47ad-8919-f245c183a8ab", "address": "fa:16:3e:d1:d0:a6", "network": {"id": "1a32c316-9a81-4aa2-a49d-80ff12af4579", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1064776770-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccad355f4d8546d2863753bfb1c86ba9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487dbf21-ef", "ovs_interfaceid": "487dbf21-efa0-47ad-8919-f245c183a8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 820.753261] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:d0:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31ac3fea-ebf4-4bed-bf70-1eaecdf71280', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '487dbf21-efa0-47ad-8919-f245c183a8ab', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.760634] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Creating folder: Project (ccad355f4d8546d2863753bfb1c86ba9). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.761550] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e801837d-7c98-45f7-b6aa-3e07ef4d2512 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.764511] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46131f06-ddac-48fd-a303-6a826c0d4261 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.796250] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742/volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.798786] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb2ffc2d-107b-4965-a2e2-c544e0c004c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.813745] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Created folder: Project (ccad355f4d8546d2863753bfb1c86ba9) in parent group-v992574. [ 820.814039] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Creating folder: Instances. Parent ref: group-v992734. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.817453] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9eb7b129-c65c-45e7-abe6-684ef184dbc1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.826241] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 820.826241] env[65107]: value = "task-5102784" [ 820.826241] env[65107]: _type = "Task" [ 820.826241] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.831649] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Created folder: Instances in parent group-v992734. [ 820.831918] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 820.832601] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.832836] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d25bd64a-195d-4903-bd66-469671cf0253 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.855997] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102784, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.862832] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.862832] env[65107]: value = "task-5102785" [ 820.862832] env[65107]: _type = "Task" [ 820.862832] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.875475] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102785, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.199718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-84a86330-025b-40fa-8925-c1cc8bcff211 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "f58c3088-c821-4eec-be0a-81221debc98e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.230s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.242342] env[65107]: DEBUG oslo_vmware.api [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102781, 'name': ReconfigVM_Task, 'duration_secs': 1.475426} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.246526] env[65107]: WARNING neutronclient.v2_0.client [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.249830] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 821.249830] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Reconfigured VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 821.342889] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102784, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.377058] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102785, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.382192] env[65107]: WARNING openstack [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.382589] env[65107]: WARNING openstack [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.629276] env[65107]: WARNING neutronclient.v2_0.client [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.629981] env[65107]: WARNING openstack [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.630347] env[65107]: WARNING openstack [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.679474] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0836f04-1867-4e0d-93cb-91605eea3ee8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.687976] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c39605-661c-439b-9125-e818237d9391 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.725134] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e57ef1-5741-4692-82b9-061996de25f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.734664] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b8986e-21df-427c-a36a-ffa9471e4831 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.749917] env[65107]: DEBUG nova.compute.provider_tree [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.752317] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2f76dbd-c567-4a7f-b9de-b895cddd86be tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-d2d0e18c-4839-416c-b363-b93996e8ba7f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.254s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.839795] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102784, 'name': ReconfigVM_Task, 'duration_secs': 0.601074} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.840345] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Reconfigured VM instance instance-0000002b to attach disk [datastore1] volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742/volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.845841] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eb196f8-f6f6-42fc-83c2-a6e28fd531d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.866217] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 821.866217] env[65107]: value = "task-5102786" [ 821.866217] env[65107]: _type = "Task" [ 821.866217] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.880253] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102785, 'name': CreateVM_Task, 'duration_secs': 0.595808} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.891854] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.894427] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102786, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.894745] env[65107]: WARNING neutronclient.v2_0.client [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.895221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.897171] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.897568] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 821.898777] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f6829d1-92a7-4cd9-b1d6-bd3989126ab9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.903957] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 821.903957] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52656d93-2138-d871-1458-434d84ae60cc" [ 821.903957] env[65107]: _type = "Task" [ 821.903957] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.914191] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52656d93-2138-d871-1458-434d84ae60cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.054970] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e82c88-9651-d913-8977-9d9b975d7788/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 822.057056] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c974a42-a400-4904-ab46-3a58d2d32f21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.063291] env[65107]: DEBUG nova.network.neutron [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updated VIF entry in instance network info cache for port 97579a1b-69c4-4b0c-b832-426f4d60592d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 822.063846] env[65107]: DEBUG nova.network.neutron [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "97579a1b-69c4-4b0c-b832-426f4d60592d", "address": "fa:16:3e:c4:9c:4c", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97579a1b-69", "ovs_interfaceid": "97579a1b-69c4-4b0c-b832-426f4d60592d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 822.067172] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e82c88-9651-d913-8977-9d9b975d7788/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 822.067459] env[65107]: ERROR oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e82c88-9651-d913-8977-9d9b975d7788/disk-0.vmdk due to incomplete transfer. [ 822.067984] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-218b5081-7f7d-4019-b758-3c6add7d1119 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.079952] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e82c88-9651-d913-8977-9d9b975d7788/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 822.080233] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Uploaded image 441569b7-9bcf-462e-b53e-b76f2b2cd9a4 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 822.083148] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 822.084921] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9e2aa482-022b-4950-a020-963f658049b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.092956] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 822.092956] env[65107]: value = "task-5102787" [ 822.092956] env[65107]: _type = "Task" [ 822.092956] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.105710] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102787, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.255673] env[65107]: DEBUG nova.scheduler.client.report [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 822.376792] env[65107]: DEBUG oslo_vmware.api [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102786, 'name': ReconfigVM_Task, 'duration_secs': 0.21867} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.377467] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992733', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'name': 'volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '82e8c093-b718-4d38-9682-ba2e710e4b93', 'attached_at': '', 'detached_at': '', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'serial': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 822.416581] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52656d93-2138-d871-1458-434d84ae60cc, 'name': SearchDatastore_Task, 'duration_secs': 0.014239} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.416929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.417483] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.417483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.417671] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.417708] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.417989] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ef7ae35-2cdf-43e0-b792-5316fa10c87c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.429721] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.429883] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.430745] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-733e9b31-2359-48de-a440-68e865b2d38d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.439020] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 822.439020] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5224a5cc-33ec-ad5c-1a30-422eab982dab" [ 822.439020] env[65107]: _type = "Task" [ 822.439020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.453421] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5224a5cc-33ec-ad5c-1a30-422eab982dab, 'name': SearchDatastore_Task, 'duration_secs': 0.01156} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.453966] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-909158ee-e9f7-4716-b1e8-fb6aef711fcc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.463762] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 822.463762] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520f8cdd-c736-9d59-ce8f-e7895660077e" [ 822.463762] env[65107]: _type = "Task" [ 822.463762] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.474368] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520f8cdd-c736-9d59-ce8f-e7895660077e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.568926] env[65107]: DEBUG oslo_concurrency.lockutils [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Releasing lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.569231] env[65107]: DEBUG nova.compute.manager [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Received event network-vif-plugged-487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 822.569677] env[65107]: DEBUG oslo_concurrency.lockutils [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Acquiring lock "5fccc802-f5ac-4450-8246-4cf9a5371046-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.569677] env[65107]: DEBUG oslo_concurrency.lockutils [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.570550] env[65107]: DEBUG oslo_concurrency.lockutils [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.570550] env[65107]: DEBUG nova.compute.manager [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] No waiting events found dispatching network-vif-plugged-487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 822.570550] env[65107]: WARNING nova.compute.manager [req-d9745f0a-beee-4352-ba39-97dba92a45fb req-567cb4de-c4d0-4700-8b89-56e3da840ab0 service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Received unexpected event network-vif-plugged-487dbf21-efa0-47ad-8919-f245c183a8ab for instance with vm_state building and task_state spawning. [ 822.613028] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102787, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.696934] env[65107]: DEBUG nova.compute.manager [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Received event network-changed-bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 822.697211] env[65107]: DEBUG nova.compute.manager [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Refreshing instance network info cache due to event network-changed-bec67e7b-de92-407d-aef5-04d4430d3ee9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 822.697384] env[65107]: DEBUG oslo_concurrency.lockutils [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Acquiring lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.697502] env[65107]: DEBUG oslo_concurrency.lockutils [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Acquired lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.697672] env[65107]: DEBUG nova.network.neutron [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Refreshing network info cache for port bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 822.765414] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.109s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.768225] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.230s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.769890] env[65107]: INFO nova.compute.claims [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.804339] env[65107]: INFO nova.scheduler.client.report [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Deleted allocations for instance 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c [ 822.975779] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520f8cdd-c736-9d59-ce8f-e7895660077e, 'name': SearchDatastore_Task, 'duration_secs': 0.018028} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.976121] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.976661] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 5fccc802-f5ac-4450-8246-4cf9a5371046/5fccc802-f5ac-4450-8246-4cf9a5371046.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 822.976863] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5f4ee16-03a1-466e-a314-cf49dacbfe9f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.986741] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 822.986741] env[65107]: value = "task-5102788" [ 822.986741] env[65107]: _type = "Task" [ 822.986741] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.996796] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.104882] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102787, 'name': Destroy_Task, 'duration_secs': 0.989306} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.105250] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroyed the VM [ 823.105438] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 823.105728] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-187fe7a9-f89f-424b-b6cd-8c5ed1c5ad21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.113849] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 823.113849] env[65107]: value = "task-5102789" [ 823.113849] env[65107]: _type = "Task" [ 823.113849] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.126252] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102789, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.204099] env[65107]: WARNING neutronclient.v2_0.client [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.204099] env[65107]: WARNING openstack [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.204825] env[65107]: WARNING openstack [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.315551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-963fec51-d8d3-46d5-8219-1d3af0fcb6bf tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.008s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.317091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 33.486s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.317091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.317313] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.317521] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.320939] env[65107]: INFO nova.compute.manager [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Terminating instance [ 823.392420] env[65107]: WARNING openstack [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.392746] env[65107]: WARNING openstack [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.407452] env[65107]: DEBUG nova.compute.manager [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Received event network-changed-487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 823.408315] env[65107]: DEBUG nova.compute.manager [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Refreshing instance network info cache due to event network-changed-487dbf21-efa0-47ad-8919-f245c183a8ab. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 823.408315] env[65107]: DEBUG oslo_concurrency.lockutils [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Acquiring lock "refresh_cache-5fccc802-f5ac-4450-8246-4cf9a5371046" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.408315] env[65107]: DEBUG oslo_concurrency.lockutils [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Acquired lock "refresh_cache-5fccc802-f5ac-4450-8246-4cf9a5371046" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 823.408447] env[65107]: DEBUG nova.network.neutron [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Refreshing network info cache for port 487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 823.410886] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "b779b5ad-7d93-47c2-b824-6d76246c00f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.410886] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.411191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "b779b5ad-7d93-47c2-b824-6d76246c00f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.411487] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.411735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.418805] env[65107]: INFO nova.compute.manager [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Terminating instance [ 823.432996] env[65107]: DEBUG nova.objects.instance [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'flavor' on Instance uuid 82e8c093-b718-4d38-9682-ba2e710e4b93 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 823.494060] env[65107]: WARNING neutronclient.v2_0.client [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.495724] env[65107]: WARNING openstack [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.495724] env[65107]: WARNING openstack [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.513034] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102788, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.609724] env[65107]: DEBUG nova.network.neutron [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updated VIF entry in instance network info cache for port bec67e7b-de92-407d-aef5-04d4430d3ee9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 823.610148] env[65107]: DEBUG nova.network.neutron [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updating instance_info_cache with network_info: [{"id": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "address": "fa:16:3e:0f:eb:48", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbec67e7b-de", "ovs_interfaceid": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 823.629943] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102789, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.824982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.825266] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquired lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 823.825374] env[65107]: DEBUG nova.network.neutron [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 823.915647] env[65107]: WARNING neutronclient.v2_0.client [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.916399] env[65107]: WARNING openstack [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.916749] env[65107]: WARNING openstack [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.925312] env[65107]: DEBUG nova.compute.manager [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 823.925556] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.926464] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35271a0-6380-45c0-ada5-f809188a4e46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.935066] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 823.935357] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c7fd38e-52f2-4590-a913-3f332c6e99a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.942177] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46c2725b-e566-44ec-bbfc-f4a7b9ce7169 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.849s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.949323] env[65107]: DEBUG oslo_vmware.api [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 823.949323] env[65107]: value = "task-5102790" [ 823.949323] env[65107]: _type = "Task" [ 823.949323] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.977970] env[65107]: DEBUG oslo_vmware.api [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102790, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.003649] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585227} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.004070] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 5fccc802-f5ac-4450-8246-4cf9a5371046/5fccc802-f5ac-4450-8246-4cf9a5371046.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.004397] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.004709] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50147473-3515-4aca-bb13-4f04ed82b665 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.018646] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 824.018646] env[65107]: value = "task-5102791" [ 824.018646] env[65107]: _type = "Task" [ 824.018646] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.030642] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102791, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.103359] env[65107]: WARNING openstack [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.104402] env[65107]: WARNING openstack [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.114481] env[65107]: DEBUG oslo_concurrency.lockutils [req-58983b87-a64b-4ac7-a3a4-7a3113a26c6d req-5a8cfdea-5b7f-43e9-b413-47c598edd52e service nova] Releasing lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.132030] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102789, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.142220] env[65107]: INFO nova.compute.manager [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Rebuilding instance [ 824.204784] env[65107]: WARNING neutronclient.v2_0.client [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.205557] env[65107]: WARNING openstack [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.206016] env[65107]: WARNING openstack [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.221086] env[65107]: DEBUG nova.compute.manager [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 824.222180] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff4debe-e943-4c24-9823-3e9249132486 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.327991] env[65107]: DEBUG nova.network.neutron [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Updated VIF entry in instance network info cache for port 487dbf21-efa0-47ad-8919-f245c183a8ab. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 824.328748] env[65107]: DEBUG nova.network.neutron [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Updating instance_info_cache with network_info: [{"id": "487dbf21-efa0-47ad-8919-f245c183a8ab", "address": "fa:16:3e:d1:d0:a6", "network": {"id": "1a32c316-9a81-4aa2-a49d-80ff12af4579", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1064776770-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccad355f4d8546d2863753bfb1c86ba9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap487dbf21-ef", "ovs_interfaceid": "487dbf21-efa0-47ad-8919-f245c183a8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 824.331124] env[65107]: DEBUG nova.compute.utils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Can not refresh info_cache because instance was not found {{(pid=65107) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 824.331489] env[65107]: WARNING neutronclient.v2_0.client [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.332133] env[65107]: WARNING openstack [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.332722] env[65107]: WARNING openstack [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.357020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3c727a-bb9b-40f8-9956-624d0b64c7ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.363693] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb324421-6b23-4145-9e2f-75d6e1cfebc8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.369896] env[65107]: DEBUG nova.network.neutron [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 824.400876] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbe46ee-c493-4708-b7f1-011f276b5bfb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.410044] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9314fd2-3e4d-4293-ab12-8c1832c0afb9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.431647] env[65107]: DEBUG nova.compute.provider_tree [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.460841] env[65107]: DEBUG oslo_vmware.api [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102790, 'name': PowerOffVM_Task, 'duration_secs': 0.341631} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.461114] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 824.461558] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 824.461558] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4f7cad9-c946-4280-bb36-71274603f421 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.476764] env[65107]: DEBUG nova.network.neutron [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 824.530724] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102791, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086935} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.530724] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.530983] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7044c515-3c68-4b33-97b0-ed74711cc601 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.558853] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 5fccc802-f5ac-4450-8246-4cf9a5371046/5fccc802-f5ac-4450-8246-4cf9a5371046.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.560035] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5eaa3f49-9e7d-4da9-866a-0ede9db010a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.575173] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 824.575397] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 824.575577] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleting the datastore file [datastore2] b779b5ad-7d93-47c2-b824-6d76246c00f5 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.575855] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61b12674-eb60-4e21-9791-e8ba1bb5ca07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.584826] env[65107]: DEBUG oslo_vmware.api [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 824.584826] env[65107]: value = "task-5102793" [ 824.584826] env[65107]: _type = "Task" [ 824.584826] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.585101] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 824.585101] env[65107]: value = "task-5102794" [ 824.585101] env[65107]: _type = "Task" [ 824.585101] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.597790] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102794, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.601511] env[65107]: DEBUG oslo_vmware.api [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.629428] env[65107]: DEBUG oslo_vmware.api [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102789, 'name': RemoveSnapshot_Task, 'duration_secs': 1.01743} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.629711] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 824.629953] env[65107]: INFO nova.compute.manager [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 16.74 seconds to snapshot the instance on the hypervisor. [ 824.806735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "ee50d08a-57fc-4c05-96fe-a11fe2708165" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.807099] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.833964] env[65107]: DEBUG oslo_concurrency.lockutils [req-7deb1d5c-e128-459e-8a54-c985d931a056 req-e8b6c4d8-6d08-42a5-9a53-fae20c0f9bcf service nova] Releasing lock "refresh_cache-5fccc802-f5ac-4450-8246-4cf9a5371046" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.937574] env[65107]: DEBUG nova.scheduler.client.report [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.979929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Releasing lock "refresh_cache-5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.981039] env[65107]: DEBUG nova.compute.manager [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 824.981277] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 824.982357] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60ccd845-a2a7-47f5-a16a-6f4e599320fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.993502] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679a0e9b-5ad3-4b88-a908-153fce048660 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.034794] env[65107]: WARNING nova.virt.vmwareapi.vmops [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c could not be found. [ 825.035373] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.035790] env[65107]: INFO nova.compute.manager [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 825.036436] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 825.036436] env[65107]: DEBUG nova.compute.manager [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 825.036701] env[65107]: DEBUG nova.network.neutron [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 825.036875] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.037532] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.037817] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.098530] env[65107]: DEBUG oslo_vmware.api [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456825} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.102321] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.102573] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 825.102816] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.103010] env[65107]: INFO nova.compute.manager [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Took 1.18 seconds to destroy the instance on the hypervisor. [ 825.103274] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 825.103529] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102794, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.103709] env[65107]: DEBUG nova.compute.manager [-] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 825.103807] env[65107]: DEBUG nova.network.neutron [-] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 825.104075] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.104615] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.104887] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.139975] env[65107]: DEBUG nova.network.neutron [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 825.140233] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.188567] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.195014] env[65107]: DEBUG nova.compute.manager [None req-e7435bfb-26b2-45cc-bb29-8f2d41530534 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Found 2 images (rotation: 2) {{(pid=65107) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 825.249793] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 825.249793] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c00fa55a-97ce-4e45-8e27-25e0899e1726 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.259482] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 825.259482] env[65107]: value = "task-5102795" [ 825.259482] env[65107]: _type = "Task" [ 825.259482] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.270067] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102795, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.289178] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 825.289678] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 825.309869] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 825.447162] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.447162] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 825.449278] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.089s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 825.451374] env[65107]: INFO nova.compute.claims [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.555573] env[65107]: DEBUG nova.compute.manager [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Received event network-changed-bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 825.555747] env[65107]: DEBUG nova.compute.manager [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Refreshing instance network info cache due to event network-changed-bec67e7b-de92-407d-aef5-04d4430d3ee9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 825.556311] env[65107]: DEBUG oslo_concurrency.lockutils [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Acquiring lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.556311] env[65107]: DEBUG oslo_concurrency.lockutils [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Acquired lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.556501] env[65107]: DEBUG nova.network.neutron [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Refreshing network info cache for port bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 825.599733] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102794, 'name': ReconfigVM_Task, 'duration_secs': 0.725679} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.600048] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 5fccc802-f5ac-4450-8246-4cf9a5371046/5fccc802-f5ac-4450-8246-4cf9a5371046.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.600936] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-994af899-ee68-43fa-b270-e0ebbc357917 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.607579] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 825.607579] env[65107]: value = "task-5102796" [ 825.607579] env[65107]: _type = "Task" [ 825.607579] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.616433] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102796, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.646944] env[65107]: DEBUG nova.network.neutron [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 825.770045] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102795, 'name': PowerOffVM_Task, 'duration_secs': 0.195929} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.770336] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 825.770572] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.771427] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1ec44e-d43e-4a77-93b6-6ed0debc7a87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.779790] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 825.780132] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-716d8730-27dd-42c2-a8a1-a216dee74e93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.791912] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 825.805519] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 825.805758] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 825.805943] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Deleting the datastore file [datastore2] 780f910f-4f73-41ba-a795-0daae3097314 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.806249] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-968ddfcc-2327-4bc6-a3ac-87e65c916b7a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.819915] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 825.819915] env[65107]: value = "task-5102798" [ 825.819915] env[65107]: _type = "Task" [ 825.819915] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.832135] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102798, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.838318] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 825.940039] env[65107]: DEBUG nova.network.neutron [-] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 825.958721] env[65107]: DEBUG nova.compute.utils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 825.962609] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 825.962898] env[65107]: DEBUG nova.network.neutron [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 825.963303] env[65107]: WARNING neutronclient.v2_0.client [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.964058] env[65107]: WARNING neutronclient.v2_0.client [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.965136] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.965136] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.040224] env[65107]: DEBUG nova.policy [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1ba49e3eedf41edbd193a027336c606', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96f4837ea5694dee83258649221accc9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 826.062558] env[65107]: WARNING neutronclient.v2_0.client [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 826.063694] env[65107]: WARNING openstack [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.064123] env[65107]: WARNING openstack [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.122428] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102796, 'name': Rename_Task, 'duration_secs': 0.165025} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.122673] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.122901] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b76dfe40-1c5b-4253-89c5-544a452d6e54 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.131098] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 826.131098] env[65107]: value = "task-5102799" [ 826.131098] env[65107]: _type = "Task" [ 826.131098] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.140677] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102799, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.151942] env[65107]: INFO nova.compute.manager [-] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Took 1.11 seconds to deallocate network for instance. [ 826.230016] env[65107]: WARNING openstack [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.230439] env[65107]: WARNING openstack [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.317369] env[65107]: WARNING neutronclient.v2_0.client [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 826.318071] env[65107]: WARNING openstack [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.318418] env[65107]: WARNING openstack [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 826.329396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.347556] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102798, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104954} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.347950] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.348999] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 826.348999] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 826.425049] env[65107]: DEBUG nova.network.neutron [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Successfully created port: 535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 826.442289] env[65107]: INFO nova.compute.manager [-] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Took 1.34 seconds to deallocate network for instance. [ 826.465870] env[65107]: DEBUG nova.network.neutron [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updated VIF entry in instance network info cache for port bec67e7b-de92-407d-aef5-04d4430d3ee9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 826.466252] env[65107]: DEBUG nova.network.neutron [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updating instance_info_cache with network_info: [{"id": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "address": "fa:16:3e:0f:eb:48", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbec67e7b-de", "ovs_interfaceid": "bec67e7b-de92-407d-aef5-04d4430d3ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.468391] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 826.541598] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cfdd88-d9a8-4ef5-841e-01a7647ead10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.550741] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e386280d-603a-4548-ad44-17f94c9a88ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.590928] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be877a1-9e48-4bf2-91bd-8abaf07d33d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.601061] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529dccac-91a8-4583-9000-59b2a9c0a2a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.619376] env[65107]: DEBUG nova.compute.provider_tree [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.645293] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102799, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.659688] env[65107]: INFO nova.compute.manager [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance disappeared during terminate [ 826.659688] env[65107]: DEBUG oslo_concurrency.lockutils [None req-11908468-cc40-4207-93e1-1df0d4e47ab6 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.342s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.718370] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-d2d0e18c-4839-416c-b363-b93996e8ba7f-97579a1b-69c4-4b0c-b832-426f4d60592d" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.718505] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-d2d0e18c-4839-416c-b363-b93996e8ba7f-97579a1b-69c4-4b0c-b832-426f4d60592d" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.719646] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "40e17da0-b459-4619-8e9e-18c3093328c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.719904] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "40e17da0-b459-4619-8e9e-18c3093328c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.952666] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.976550] env[65107]: DEBUG oslo_concurrency.lockutils [req-bd9c7e44-aaea-44f3-b0ab-8834869b23f2 req-7159d1b9-8baa-482a-ae12-841b36398dfe service nova] Releasing lock "refresh_cache-179e613b-e5a2-4fbc-8fa6-b72769425ff3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 827.122704] env[65107]: DEBUG nova.scheduler.client.report [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 827.143693] env[65107]: DEBUG oslo_vmware.api [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102799, 'name': PowerOnVM_Task, 'duration_secs': 0.540176} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.143972] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.144200] env[65107]: INFO nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Took 8.85 seconds to spawn the instance on the hypervisor. [ 827.144378] env[65107]: DEBUG nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 827.145174] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e09303d-15b2-49da-bf5b-a687531e8a66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.220828] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.221061] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 827.222321] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0392a318-4aa8-4daf-bc95-b2a2cb01ca53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.240323] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b83873-c27c-4be6-bcc6-28e0ec5db9ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.263272] env[65107]: WARNING neutronclient.v2_0.client [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 827.269637] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Reconfiguring VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 827.270116] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a9a4b96-56c1-404b-8f2f-d156099cd64a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.297518] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 827.297518] env[65107]: value = "task-5102800" [ 827.297518] env[65107]: _type = "Task" [ 827.297518] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.307988] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.339305] env[65107]: DEBUG nova.compute.manager [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 827.340366] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca1bd35-8341-44d9-9a71-0929ea0aaee4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.364933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.364933] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.365143] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.365408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.365408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.368199] env[65107]: INFO nova.compute.manager [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Terminating instance [ 827.393533] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 827.393836] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 827.393955] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 827.394160] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 827.394308] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 827.394452] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 827.394656] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.394813] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 827.394974] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 827.395157] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 827.395330] env[65107]: DEBUG nova.virt.hardware [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 827.396049] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3f34d6a2-0370-4239-9556-b4333e454d58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.396276] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3f34d6a2-0370-4239-9556-b4333e454d58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.397319] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbdf517-ecff-46e7-b491-47c6afe244c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.407658] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e7f258-ad96-4079-828b-60fc7b488ce6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.425983] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.431820] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 827.432782] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.432782] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1084b3a-d272-4f96-81dd-8e84e7a75dfa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.451813] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.451813] env[65107]: value = "task-5102801" [ 827.451813] env[65107]: _type = "Task" [ 827.451813] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.461660] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102801, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.485840] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 827.510494] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 827.510738] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 827.510897] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 827.511233] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 827.511507] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 827.511575] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 827.511744] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.511898] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 827.512077] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 827.512244] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 827.512449] env[65107]: DEBUG nova.virt.hardware [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 827.513995] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19af4122-0019-4fb3-b0c3-67541b79e8ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.523103] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc69d860-79b2-423e-856f-b24921d50853 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.628318] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.179s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.628888] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 827.631822] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.051s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.632168] env[65107]: DEBUG nova.objects.instance [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lazy-loading 'resources' on Instance uuid 19518d90-ba6f-44e5-b7c7-1ececad2b54d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.663754] env[65107]: INFO nova.compute.manager [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Took 51.12 seconds to build instance. [ 827.744338] env[65107]: DEBUG nova.compute.manager [req-2eb02de2-55e3-469d-ab5d-da1d7b32c3dd req-dad8c414-6a3c-4c86-9227-e4933c859467 service nova] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Received event network-vif-deleted-3ce4a5c1-5dae-42ba-919f-736e66893243 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 827.808292] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.854139] env[65107]: INFO nova.compute.manager [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] instance snapshotting [ 827.854812] env[65107]: DEBUG nova.objects.instance [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'flavor' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.873965] env[65107]: DEBUG nova.compute.manager [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 827.874557] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.875198] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0853292-58d9-4023-bed1-014bbec33d5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.886040] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 827.886314] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34288418-f6ad-489d-926d-26fcd5a53690 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.894104] env[65107]: DEBUG oslo_vmware.api [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 827.894104] env[65107]: value = "task-5102802" [ 827.894104] env[65107]: _type = "Task" [ 827.894104] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.905319] env[65107]: DEBUG oslo_vmware.api [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.967854] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102801, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.986113] env[65107]: DEBUG nova.network.neutron [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Successfully updated port: 535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 828.139137] env[65107]: DEBUG nova.compute.utils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 828.140610] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 828.140832] env[65107]: DEBUG nova.network.neutron [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 828.141330] env[65107]: WARNING neutronclient.v2_0.client [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.141576] env[65107]: WARNING neutronclient.v2_0.client [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.142384] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.142870] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.166710] env[65107]: DEBUG oslo_concurrency.lockutils [None req-74ddd1be-98b8-4f8e-a5e4-da01d4d7a94b tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.632s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.195838] env[65107]: DEBUG nova.policy [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32155078a41249a297470fd4083860df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f9188f3e35e4d06a977e2180918c616', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 828.315456] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.361841] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2ca6b8-f5da-499a-9934-0fcb29d846e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.388306] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18387605-acd3-42da-acf1-e0ce828d7fbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.416114] env[65107]: DEBUG oslo_vmware.api [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102802, 'name': PowerOffVM_Task, 'duration_secs': 0.196947} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.417455] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 828.417455] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 828.417455] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1619bee5-d423-48c6-a98a-6e0b4266b56f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.466358] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102801, 'name': CreateVM_Task, 'duration_secs': 0.586216} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.466492] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.467040] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.467209] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.467826] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 828.467826] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d81cea6-4c34-4d6b-b411-09dce8d19d23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.477170] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 828.477170] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d8a123-1271-27d5-541d-527bd2fbd37c" [ 828.477170] env[65107]: _type = "Task" [ 828.477170] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.488121] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "refresh_cache-805e2210-fb58-48da-be63-5d548b2f8a2f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.489342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquired lock "refresh_cache-805e2210-fb58-48da-be63-5d548b2f8a2f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.489342] env[65107]: DEBUG nova.network.neutron [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 828.490757] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d8a123-1271-27d5-541d-527bd2fbd37c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.495461] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 828.496117] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 828.496117] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Deleting the datastore file [datastore1] ef417f32-8c0e-4a62-9078-cb2bf95b85f8 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.496117] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f7c55d6-93a0-498d-a9bb-d4523f99e15f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.504489] env[65107]: DEBUG oslo_vmware.api [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for the task: (returnval){ [ 828.504489] env[65107]: value = "task-5102804" [ 828.504489] env[65107]: _type = "Task" [ 828.504489] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.518786] env[65107]: DEBUG nova.network.neutron [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Successfully created port: acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 828.526301] env[65107]: DEBUG oslo_vmware.api [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102804, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.650639] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 828.672686] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 828.716849] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee144695-97c7-4d94-ba11-a7475bfe066f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.726941] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3332473-1aee-4c53-ae0f-378623c6782f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.763390] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6310e7-2850-4db4-89fa-0e405d1724d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.771897] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5571d5de-3e7a-4734-a364-89912abb75b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.787009] env[65107]: DEBUG nova.compute.provider_tree [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.808479] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.908026] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 828.908026] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e864f94e-b877-403c-a1af-7cfb7d5c8c12 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.916521] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 828.916521] env[65107]: value = "task-5102805" [ 828.916521] env[65107]: _type = "Task" [ 828.916521] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.926669] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.988930] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d8a123-1271-27d5-541d-527bd2fbd37c, 'name': SearchDatastore_Task, 'duration_secs': 0.011245} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.989685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.989774] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.990174] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.990515] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.990515] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.990765] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a8590dc-1e81-4700-a144-4f00a4dba34d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.995044] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.995484] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.004273] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.004809] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.005546] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-318f0da2-28c8-4777-912d-a37e037c73d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.016758] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 829.016758] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528a72f5-7d93-41ae-8049-61498f7a26f5" [ 829.016758] env[65107]: _type = "Task" [ 829.016758] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.020640] env[65107]: DEBUG oslo_vmware.api [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Task: {'id': task-5102804, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157362} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.023919] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.024150] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.024333] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.024506] env[65107]: INFO nova.compute.manager [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 829.024778] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 829.026383] env[65107]: DEBUG nova.compute.manager [-] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 829.026383] env[65107]: DEBUG nova.network.neutron [-] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 829.026383] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.026383] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.026383] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.040306] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528a72f5-7d93-41ae-8049-61498f7a26f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010813} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.040842] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-442c8032-1642-4f96-9902-5b994be9616e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.046769] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 829.046769] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527414ed-4892-e653-fc66-13ef1002447e" [ 829.046769] env[65107]: _type = "Task" [ 829.046769] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.047802] env[65107]: DEBUG nova.network.neutron [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 829.058617] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527414ed-4892-e653-fc66-13ef1002447e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.105927] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.189408] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.189841] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.197803] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.290953] env[65107]: DEBUG nova.scheduler.client.report [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 829.313599] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.432329] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.545662] env[65107]: WARNING neutronclient.v2_0.client [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.546353] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.546719] env[65107]: WARNING openstack [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.565376] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527414ed-4892-e653-fc66-13ef1002447e, 'name': SearchDatastore_Task, 'duration_secs': 0.011662} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.566038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.566038] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.566197] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93c8e719-fe95-4cc9-b5a3-471b6374c2ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.574925] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 829.574925] env[65107]: value = "task-5102806" [ 829.574925] env[65107]: _type = "Task" [ 829.574925] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.584221] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.660376] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 829.693019] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 829.693807] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 829.694143] env[65107]: DEBUG nova.virt.hardware [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 829.695104] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28f2a2c-1ae3-4931-8f97-289e7701d274 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.704604] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacdc468-4d07-43eb-ad1b-c2331cd8850e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.788687] env[65107]: DEBUG nova.network.neutron [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Updating instance_info_cache with network_info: [{"id": "535ab747-8154-4faf-ae33-4f3a850a5dfb", "address": "fa:16:3e:1b:d5:2d", "network": {"id": "1240b27b-e287-443a-ba3d-38655cc74faf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1118025022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96f4837ea5694dee83258649221accc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap535ab747-81", "ovs_interfaceid": "535ab747-8154-4faf-ae33-4f3a850a5dfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 829.797038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.164s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.798539] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.340s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.801464] env[65107]: INFO nova.compute.claims [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.816944] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.818413] env[65107]: INFO nova.scheduler.client.report [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted allocations for instance 19518d90-ba6f-44e5-b7c7-1ececad2b54d [ 829.899264] env[65107]: DEBUG nova.network.neutron [-] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 829.929202] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102805, 'name': CreateSnapshot_Task, 'duration_secs': 0.628624} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.929664] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 829.930479] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1377c6ff-9353-4d82-a43b-fdb23300be65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.087986] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102806, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.160292] env[65107]: DEBUG nova.network.neutron [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Successfully updated port: acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 830.258023] env[65107]: DEBUG nova.compute.manager [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Received event network-vif-plugged-535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 830.258253] env[65107]: DEBUG oslo_concurrency.lockutils [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Acquiring lock "805e2210-fb58-48da-be63-5d548b2f8a2f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.258461] env[65107]: DEBUG oslo_concurrency.lockutils [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.258624] env[65107]: DEBUG oslo_concurrency.lockutils [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.258834] env[65107]: DEBUG nova.compute.manager [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] No waiting events found dispatching network-vif-plugged-535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 830.258982] env[65107]: WARNING nova.compute.manager [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Received unexpected event network-vif-plugged-535ab747-8154-4faf-ae33-4f3a850a5dfb for instance with vm_state building and task_state spawning. [ 830.259158] env[65107]: DEBUG nova.compute.manager [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Received event network-changed-535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 830.259308] env[65107]: DEBUG nova.compute.manager [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Refreshing instance network info cache due to event network-changed-535ab747-8154-4faf-ae33-4f3a850a5dfb. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 830.259464] env[65107]: DEBUG oslo_concurrency.lockutils [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Acquiring lock "refresh_cache-805e2210-fb58-48da-be63-5d548b2f8a2f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.291729] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Releasing lock "refresh_cache-805e2210-fb58-48da-be63-5d548b2f8a2f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.292151] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Instance network_info: |[{"id": "535ab747-8154-4faf-ae33-4f3a850a5dfb", "address": "fa:16:3e:1b:d5:2d", "network": {"id": "1240b27b-e287-443a-ba3d-38655cc74faf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1118025022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96f4837ea5694dee83258649221accc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap535ab747-81", "ovs_interfaceid": "535ab747-8154-4faf-ae33-4f3a850a5dfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 830.292892] env[65107]: DEBUG oslo_concurrency.lockutils [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Acquired lock "refresh_cache-805e2210-fb58-48da-be63-5d548b2f8a2f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.293087] env[65107]: DEBUG nova.network.neutron [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Refreshing network info cache for port 535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 830.294771] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:d5:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '821e0800-fabf-48d0-87b4-db5a1eddce93', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '535ab747-8154-4faf-ae33-4f3a850a5dfb', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.303481] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Creating folder: Project (96f4837ea5694dee83258649221accc9). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.304893] env[65107]: WARNING neutronclient.v2_0.client [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.305678] env[65107]: WARNING openstack [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.306039] env[65107]: WARNING openstack [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.316099] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d774f9f7-dcaf-43d3-a4a8-9b51517efbd3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.333107] env[65107]: DEBUG oslo_concurrency.lockutils [None req-df446457-d5da-4cad-94a2-3eb662dfddcb tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "19518d90-ba6f-44e5-b7c7-1ececad2b54d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.869s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.338589] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.340468] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Created folder: Project (96f4837ea5694dee83258649221accc9) in parent group-v992574. [ 830.340733] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Creating folder: Instances. Parent ref: group-v992739. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.340937] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0f55478-7700-42f5-916b-fd24a360c91f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.353205] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Created folder: Instances in parent group-v992739. [ 830.353402] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 830.354411] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.354411] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d469e6f7-af56-4fc0-a488-23f19b1943d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.379084] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.379084] env[65107]: value = "task-5102809" [ 830.379084] env[65107]: _type = "Task" [ 830.379084] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.388515] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102809, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.403069] env[65107]: INFO nova.compute.manager [-] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Took 1.38 seconds to deallocate network for instance. [ 830.410318] env[65107]: DEBUG nova.compute.manager [req-6c1e2397-3eb9-4d38-9d6f-5b9cd7753876 req-9c774dae-f210-4dc8-8776-b18b57813bb5 service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Received event network-vif-plugged-acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 830.410318] env[65107]: DEBUG oslo_concurrency.lockutils [req-6c1e2397-3eb9-4d38-9d6f-5b9cd7753876 req-9c774dae-f210-4dc8-8776-b18b57813bb5 service nova] Acquiring lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.410318] env[65107]: DEBUG oslo_concurrency.lockutils [req-6c1e2397-3eb9-4d38-9d6f-5b9cd7753876 req-9c774dae-f210-4dc8-8776-b18b57813bb5 service nova] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.410523] env[65107]: DEBUG oslo_concurrency.lockutils [req-6c1e2397-3eb9-4d38-9d6f-5b9cd7753876 req-9c774dae-f210-4dc8-8776-b18b57813bb5 service nova] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.410622] env[65107]: DEBUG nova.compute.manager [req-6c1e2397-3eb9-4d38-9d6f-5b9cd7753876 req-9c774dae-f210-4dc8-8776-b18b57813bb5 service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] No waiting events found dispatching network-vif-plugged-acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 830.410733] env[65107]: WARNING nova.compute.manager [req-6c1e2397-3eb9-4d38-9d6f-5b9cd7753876 req-9c774dae-f210-4dc8-8776-b18b57813bb5 service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Received unexpected event network-vif-plugged-acc9d6b8-be91-4755-a23c-2c530ae497fe for instance with vm_state building and task_state spawning. [ 830.444169] env[65107]: WARNING openstack [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.444570] env[65107]: WARNING openstack [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.459056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 830.460401] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c4f2b8c0-a25d-45b3-8ac2-bbd48b3f1e2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.469621] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 830.469621] env[65107]: value = "task-5102810" [ 830.469621] env[65107]: _type = "Task" [ 830.469621] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.479596] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102810, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.521352] env[65107]: WARNING neutronclient.v2_0.client [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.522067] env[65107]: WARNING openstack [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.522437] env[65107]: WARNING openstack [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.588227] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605434} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.588600] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.588935] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.589394] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6aca230f-afdc-459e-aa55-84bbd3867202 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.599103] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 830.599103] env[65107]: value = "task-5102811" [ 830.599103] env[65107]: _type = "Task" [ 830.599103] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.609972] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.624237] env[65107]: DEBUG nova.network.neutron [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Updated VIF entry in instance network info cache for port 535ab747-8154-4faf-ae33-4f3a850a5dfb. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 830.624649] env[65107]: DEBUG nova.network.neutron [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Updating instance_info_cache with network_info: [{"id": "535ab747-8154-4faf-ae33-4f3a850a5dfb", "address": "fa:16:3e:1b:d5:2d", "network": {"id": "1240b27b-e287-443a-ba3d-38655cc74faf", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1118025022-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "96f4837ea5694dee83258649221accc9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "821e0800-fabf-48d0-87b4-db5a1eddce93", "external-id": "nsx-vlan-transportzone-97", "segmentation_id": 97, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap535ab747-81", "ovs_interfaceid": "535ab747-8154-4faf-ae33-4f3a850a5dfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.662662] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "refresh_cache-2d0a93a4-59a8-4788-93ce-c34af1432a4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.662930] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "refresh_cache-2d0a93a4-59a8-4788-93ce-c34af1432a4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.663166] env[65107]: DEBUG nova.network.neutron [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 830.832169] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.895027] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102809, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.910235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.984111] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102810, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.113315] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077237} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.113712] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.114975] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54c9796-1994-4312-89c9-44d665a196a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.133419] env[65107]: DEBUG oslo_concurrency.lockutils [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] Releasing lock "refresh_cache-805e2210-fb58-48da-be63-5d548b2f8a2f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.133794] env[65107]: DEBUG nova.compute.manager [req-e8d39a7b-f1f0-41f2-ba3c-9d7993437878 req-fc5ae497-cd9c-445e-8957-416132a9ac3b service nova] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Received event network-vif-deleted-2ec2fffa-a684-4c3d-8ead-2c8b74d6252d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 831.143865] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.147059] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b31717a8-868a-46b0-a3e7-75057a398658 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.166468] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.166959] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.180169] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 831.180169] env[65107]: value = "task-5102812" [ 831.180169] env[65107]: _type = "Task" [ 831.180169] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.191432] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102812, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.217721] env[65107]: DEBUG nova.network.neutron [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 831.239376] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.239765] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.324303] env[65107]: WARNING neutronclient.v2_0.client [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.325021] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.325318] env[65107]: WARNING openstack [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.341870] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.378607] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a434b3f8-f499-4a6d-9177-c5d744e2af17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.390289] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef01c9ad-7b21-4a85-a961-9869bc926b55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.400718] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102809, 'name': CreateVM_Task, 'duration_secs': 0.759126} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.402169] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.403031] env[65107]: WARNING neutronclient.v2_0.client [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.403136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.403304] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.403631] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 831.403913] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e907ccdd-6169-4832-80ba-1c2045f3ab03 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.432339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0f7461-ff0c-45bc-86da-f123828bad53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.436871] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 831.436871] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5256f706-767a-1dc4-d97d-ce972f383458" [ 831.436871] env[65107]: _type = "Task" [ 831.436871] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.444546] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857cf981-a753-4e58-aed8-5e0ba4367293 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.451340] env[65107]: DEBUG nova.network.neutron [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Updating instance_info_cache with network_info: [{"id": "acc9d6b8-be91-4755-a23c-2c530ae497fe", "address": "fa:16:3e:bb:e9:a8", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacc9d6b8-be", "ovs_interfaceid": "acc9d6b8-be91-4755-a23c-2c530ae497fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 831.456887] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5256f706-767a-1dc4-d97d-ce972f383458, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.473251] env[65107]: DEBUG nova.compute.provider_tree [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.486999] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102810, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.691399] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102812, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.829081] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.948900] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5256f706-767a-1dc4-d97d-ce972f383458, 'name': SearchDatastore_Task, 'duration_secs': 0.035958} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.949374] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.949635] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.949877] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.950067] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.950315] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.950681] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2610e779-2583-4bad-983f-16db175eeb12 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.957920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "refresh_cache-2d0a93a4-59a8-4788-93ce-c34af1432a4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.959133] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Instance network_info: |[{"id": "acc9d6b8-be91-4755-a23c-2c530ae497fe", "address": "fa:16:3e:bb:e9:a8", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacc9d6b8-be", "ovs_interfaceid": "acc9d6b8-be91-4755-a23c-2c530ae497fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 831.959573] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:e9:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'acc9d6b8-be91-4755-a23c-2c530ae497fe', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.968026] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 831.971093] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.971093] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.971093] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.971093] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67d47725-0323-4ed2-8b94-d9c1082b6786 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.989090] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-646fedee-8e67-495f-821f-338a63094a5a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.995020] env[65107]: DEBUG nova.scheduler.client.report [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.006029] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 832.006029] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5213044f-181b-7364-65c2-02189c855a08" [ 832.006029] env[65107]: _type = "Task" [ 832.006029] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.012071] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102810, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.012314] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.012314] env[65107]: value = "task-5102813" [ 832.012314] env[65107]: _type = "Task" [ 832.012314] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.025667] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5213044f-181b-7364-65c2-02189c855a08, 'name': SearchDatastore_Task, 'duration_secs': 0.013037} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.030130] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102813, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.030381] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-472908f0-72f3-40fd-ad90-ebb530762a80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.036509] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 832.036509] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b62f2a-7b4c-aca1-f069-1155ecbcb0cf" [ 832.036509] env[65107]: _type = "Task" [ 832.036509] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.047165] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b62f2a-7b4c-aca1-f069-1155ecbcb0cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.190519] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102812, 'name': ReconfigVM_Task, 'duration_secs': 0.755145} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.190806] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 780f910f-4f73-41ba-a795-0daae3097314/780f910f-4f73-41ba-a795-0daae3097314.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.191520] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14989655-4061-4bc8-9b18-cc8a3ea661aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.200489] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 832.200489] env[65107]: value = "task-5102814" [ 832.200489] env[65107]: _type = "Task" [ 832.200489] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.211085] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102814, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.334023] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.341601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.341922] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.506986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.708s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 832.507787] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 832.511943] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102810, 'name': CloneVM_Task, 'duration_secs': 1.776192} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.512471] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.253s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.513981] env[65107]: INFO nova.compute.claims [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.517375] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Created linked-clone VM from snapshot [ 832.517767] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1cdb86-cc14-47f5-9c10-3d47887ca652 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.534096] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Uploading image 9b846d61-6f7a-470d-bef4-76e583435ca3 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 832.535791] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102813, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.549896] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b62f2a-7b4c-aca1-f069-1155ecbcb0cf, 'name': SearchDatastore_Task, 'duration_secs': 0.010219} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.550346] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.550646] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 805e2210-fb58-48da-be63-5d548b2f8a2f/805e2210-fb58-48da-be63-5d548b2f8a2f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.551224] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-608a24f1-86ce-4f2c-8885-7c319bd60236 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.560435] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 832.560435] env[65107]: value = "task-5102815" [ 832.560435] env[65107]: _type = "Task" [ 832.560435] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.570862] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102815, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.573061] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 832.573061] env[65107]: value = "vm-992742" [ 832.573061] env[65107]: _type = "VirtualMachine" [ 832.573061] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 832.573329] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3953d3a8-5b65-4a3d-b421-a3ea125165e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.578844] env[65107]: DEBUG nova.compute.manager [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Received event network-changed-acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 832.578844] env[65107]: DEBUG nova.compute.manager [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Refreshing instance network info cache due to event network-changed-acc9d6b8-be91-4755-a23c-2c530ae497fe. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 832.578844] env[65107]: DEBUG oslo_concurrency.lockutils [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Acquiring lock "refresh_cache-2d0a93a4-59a8-4788-93ce-c34af1432a4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.578844] env[65107]: DEBUG oslo_concurrency.lockutils [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Acquired lock "refresh_cache-2d0a93a4-59a8-4788-93ce-c34af1432a4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 832.578844] env[65107]: DEBUG nova.network.neutron [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Refreshing network info cache for port acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 832.585326] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease: (returnval){ [ 832.585326] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52002e5a-62e4-fed6-34ee-d186a36c9f8e" [ 832.585326] env[65107]: _type = "HttpNfcLease" [ 832.585326] env[65107]: } obtained for exporting VM: (result){ [ 832.585326] env[65107]: value = "vm-992742" [ 832.585326] env[65107]: _type = "VirtualMachine" [ 832.585326] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 832.585326] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the lease: (returnval){ [ 832.585326] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52002e5a-62e4-fed6-34ee-d186a36c9f8e" [ 832.585326] env[65107]: _type = "HttpNfcLease" [ 832.585326] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 832.593034] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 832.593034] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52002e5a-62e4-fed6-34ee-d186a36c9f8e" [ 832.593034] env[65107]: _type = "HttpNfcLease" [ 832.593034] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 832.713594] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102814, 'name': Rename_Task, 'duration_secs': 0.363027} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.714096] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.714445] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f9aac68-ffee-482e-802c-2c8d95e10e5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.723737] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 832.723737] env[65107]: value = "task-5102817" [ 832.723737] env[65107]: _type = "Task" [ 832.723737] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.735339] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.832301] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.841383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "5fccc802-f5ac-4450-8246-4cf9a5371046" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.841634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.842062] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "5fccc802-f5ac-4450-8246-4cf9a5371046-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.842232] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.842426] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 832.844764] env[65107]: INFO nova.compute.manager [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Terminating instance [ 833.014473] env[65107]: DEBUG nova.compute.utils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 833.016499] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 833.017220] env[65107]: DEBUG nova.network.neutron [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 833.017220] env[65107]: WARNING neutronclient.v2_0.client [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.017460] env[65107]: WARNING neutronclient.v2_0.client [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.018268] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.018797] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.048258] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102813, 'name': CreateVM_Task, 'duration_secs': 0.627535} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.048258] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.048420] env[65107]: WARNING neutronclient.v2_0.client [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.057797] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.057797] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.057797] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 833.057797] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65839331-830c-40a0-9850-d7dc9321d7d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.072916] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 833.072916] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528ee543-8523-d9cb-f878-bd096aea7bf9" [ 833.072916] env[65107]: _type = "Task" [ 833.072916] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.084842] env[65107]: WARNING neutronclient.v2_0.client [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.085692] env[65107]: WARNING openstack [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.086115] env[65107]: WARNING openstack [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.094830] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102815, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.105083] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528ee543-8523-d9cb-f878-bd096aea7bf9, 'name': SearchDatastore_Task} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.105894] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.106266] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.106530] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.106698] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.106864] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.107267] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7beb7c6d-4201-42d3-b1b1-e90d536be75f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.110956] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 833.110956] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52002e5a-62e4-fed6-34ee-d186a36c9f8e" [ 833.110956] env[65107]: _type = "HttpNfcLease" [ 833.110956] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 833.114343] env[65107]: DEBUG nova.policy [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '513e39583cc4477c828531bb3250de31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f883a2e25da34a548b020184acb9245e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 833.116549] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 833.116549] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52002e5a-62e4-fed6-34ee-d186a36c9f8e" [ 833.116549] env[65107]: _type = "HttpNfcLease" [ 833.116549] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 833.117603] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfac0faf-6c3d-44fd-af7e-8323f5ee3dc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.126526] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52723b0e-d7aa-b1fb-edde-3f9783ca3533/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 833.126716] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52723b0e-d7aa-b1fb-edde-3f9783ca3533/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 833.129841] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.129841] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.131322] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82843699-f8fb-484a-a81d-3bdbafd7f8d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.202503] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 833.202503] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b961dd-2608-3748-95f4-8d46fc746a2e" [ 833.202503] env[65107]: _type = "Task" [ 833.202503] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.212122] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b961dd-2608-3748-95f4-8d46fc746a2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.233157] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102817, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.257870] env[65107]: WARNING openstack [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.258304] env[65107]: WARNING openstack [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.323073] env[65107]: WARNING neutronclient.v2_0.client [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.323810] env[65107]: WARNING openstack [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.324203] env[65107]: WARNING openstack [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.341460] env[65107]: DEBUG oslo_vmware.api [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102800, 'name': ReconfigVM_Task, 'duration_secs': 5.852676} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.341714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.341922] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Reconfigured VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 833.342353] env[65107]: WARNING neutronclient.v2_0.client [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.342658] env[65107]: WARNING neutronclient.v2_0.client [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.343227] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.343584] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.350760] env[65107]: DEBUG nova.compute.manager [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 833.350954] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.351775] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0073750-6d3e-452d-9567-c845c60ed297 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.359539] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.359668] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63490396-6d57-4436-a4a1-8df97a3ee2c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.365844] env[65107]: DEBUG oslo_vmware.api [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 833.365844] env[65107]: value = "task-5102818" [ 833.365844] env[65107]: _type = "Task" [ 833.365844] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.379493] env[65107]: DEBUG oslo_vmware.api [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.383187] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-40675a4d-d1df-4ed8-bd2f-b536a2a0e16f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.434234] env[65107]: DEBUG nova.network.neutron [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Successfully created port: 860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 833.465405] env[65107]: DEBUG nova.network.neutron [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Updated VIF entry in instance network info cache for port acc9d6b8-be91-4755-a23c-2c530ae497fe. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 833.465783] env[65107]: DEBUG nova.network.neutron [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Updating instance_info_cache with network_info: [{"id": "acc9d6b8-be91-4755-a23c-2c530ae497fe", "address": "fa:16:3e:bb:e9:a8", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacc9d6b8-be", "ovs_interfaceid": "acc9d6b8-be91-4755-a23c-2c530ae497fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 833.528693] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 833.579174] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102815, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.720134] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b961dd-2608-3748-95f4-8d46fc746a2e, 'name': SearchDatastore_Task, 'duration_secs': 0.011699} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.721153] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dca4aca7-8724-410d-b1de-13142099f72f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.734146] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 833.734146] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eed0c4-f2ef-429d-e882-82b8fa682ca3" [ 833.734146] env[65107]: _type = "Task" [ 833.734146] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.742454] env[65107]: DEBUG oslo_vmware.api [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102817, 'name': PowerOnVM_Task, 'duration_secs': 0.933889} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.743420] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.743692] env[65107]: DEBUG nova.compute.manager [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 833.744693] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4271143a-15d7-4d16-905b-3aeb888737d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.751650] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eed0c4-f2ef-429d-e882-82b8fa682ca3, 'name': SearchDatastore_Task, 'duration_secs': 0.011952} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.754433] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.755223] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2d0a93a4-59a8-4788-93ce-c34af1432a4f/2d0a93a4-59a8-4788-93ce-c34af1432a4f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.759289] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fcde776-6498-4148-9cc9-fb9fe79476fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.774770] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 833.774770] env[65107]: value = "task-5102819" [ 833.774770] env[65107]: _type = "Task" [ 833.774770] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.790980] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102819, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.881507] env[65107]: DEBUG oslo_vmware.api [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102818, 'name': PowerOffVM_Task, 'duration_secs': 0.365403} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.881899] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.882079] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.882371] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37f63e45-8aaf-4a10-8447-21122e640438 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.959383] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.959581] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.959859] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Deleting the datastore file [datastore1] 5fccc802-f5ac-4450-8246-4cf9a5371046 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.960217] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4a713a2-d349-44d1-a269-6004f9b8ebab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.972040] env[65107]: DEBUG oslo_vmware.api [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for the task: (returnval){ [ 833.972040] env[65107]: value = "task-5102821" [ 833.972040] env[65107]: _type = "Task" [ 833.972040] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.975751] env[65107]: DEBUG oslo_concurrency.lockutils [req-2944c3e6-4891-45a4-a02c-b56b633131c1 req-6994e1c7-093c-46ab-b7e9-63ce51ee014b service nova] Releasing lock "refresh_cache-2d0a93a4-59a8-4788-93ce-c34af1432a4f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.985781] env[65107]: DEBUG oslo_vmware.api [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102821, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.018173] env[65107]: DEBUG nova.compute.manager [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-vif-deleted-97579a1b-69c4-4b0c-b832-426f4d60592d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 834.020064] env[65107]: INFO nova.compute.manager [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Neutron deleted interface 97579a1b-69c4-4b0c-b832-426f4d60592d; detaching it from the instance and deleting it from the info cache [ 834.020064] env[65107]: DEBUG nova.network.neutron [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 834.073795] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102815, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.147838] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e84ca39-15f6-48cf-9974-212c50b14de1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.156607] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1539cdb6-2f5c-4a2d-bac9-66a88950d09f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.197944] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d521ab03-30f5-4f7f-9794-b81d7afe7c07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.207898] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b25820-eed1-4cc7-86e2-0c7e1ddbdfdd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.226253] env[65107]: DEBUG nova.compute.provider_tree [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.278191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.288409] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102819, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.483545] env[65107]: DEBUG oslo_vmware.api [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Task: {'id': task-5102821, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204396} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.483846] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.484070] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.484253] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.484392] env[65107]: INFO nova.compute.manager [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Took 1.13 seconds to destroy the instance on the hypervisor. [ 834.484654] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 834.484889] env[65107]: DEBUG nova.compute.manager [-] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 834.485014] env[65107]: DEBUG nova.network.neutron [-] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 834.485286] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 834.485849] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 834.486218] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 834.525970] env[65107]: DEBUG oslo_concurrency.lockutils [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.526139] env[65107]: DEBUG oslo_concurrency.lockutils [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] Acquired lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.527323] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa65dde6-a997-4679-96d8-fd5cce43c22f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.550326] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 834.553781] env[65107]: DEBUG oslo_concurrency.lockutils [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] Releasing lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.553781] env[65107]: WARNING nova.compute.manager [req-e9fb86c9-04c5-48a0-93f5-78013705e4fd req-c8e0e5e2-d585-4cd1-9cd7-3a2b450b76cc service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Detach interface failed, port_id=97579a1b-69c4-4b0c-b832-426f4d60592d, reason: No device with interface-id 97579a1b-69c4-4b0c-b832-426f4d60592d exists on VM: nova.exception.NotFound: No device with interface-id 97579a1b-69c4-4b0c-b832-426f4d60592d exists on VM [ 834.556245] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 834.576593] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102815, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.55969} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.576913] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 805e2210-fb58-48da-be63-5d548b2f8a2f/805e2210-fb58-48da-be63-5d548b2f8a2f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.577263] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.580029] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 834.580403] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 834.580726] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 834.580982] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 834.581204] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 834.581470] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 834.581757] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.582047] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 834.582327] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 834.582585] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 834.583019] env[65107]: DEBUG nova.virt.hardware [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 834.583345] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84feaf2a-b0f5-417b-a64f-0b5a1b51a324 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.588020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3b34cf-5529-4a6d-8e93-7cec8cf9ef67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.598719] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89794ff-4603-47ef-a07d-f006f092488c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.605306] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 834.605306] env[65107]: value = "task-5102822" [ 834.605306] env[65107]: _type = "Task" [ 834.605306] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.625601] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102822, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.702101] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.702252] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.702637] env[65107]: DEBUG nova.network.neutron [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 834.729972] env[65107]: DEBUG nova.scheduler.client.report [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 834.788558] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102819, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.798496} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.788852] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2d0a93a4-59a8-4788-93ce-c34af1432a4f/2d0a93a4-59a8-4788-93ce-c34af1432a4f.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.789220] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.789568] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-125cf976-e577-4809-8375-e8f753489b6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.797370] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 834.797370] env[65107]: value = "task-5102823" [ 834.797370] env[65107]: _type = "Task" [ 834.797370] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.813206] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.878639] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.879189] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 834.879423] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.879511] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 834.879732] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 834.882865] env[65107]: INFO nova.compute.manager [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Terminating instance [ 835.117068] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102822, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066639} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.117475] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.118616] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b124ea2-3ff6-4763-8a53-ca1a43e330cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.143368] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 805e2210-fb58-48da-be63-5d548b2f8a2f/805e2210-fb58-48da-be63-5d548b2f8a2f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.143757] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb4f215c-c968-42bc-8992-1107f72ffc59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.161112] env[65107]: DEBUG nova.network.neutron [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Successfully updated port: 860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 835.169831] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 835.169831] env[65107]: value = "task-5102824" [ 835.169831] env[65107]: _type = "Task" [ 835.169831] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.179595] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102824, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.206029] env[65107]: WARNING neutronclient.v2_0.client [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.207209] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.207407] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.237494] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.725s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.238246] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 835.241425] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.089s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.243290] env[65107]: INFO nova.compute.claims [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.305221] env[65107]: DEBUG nova.network.neutron [-] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 835.313405] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076284} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.319159] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.320762] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96579842-2907-4988-8e8d-5bfb30d0ea98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.348112] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 2d0a93a4-59a8-4788-93ce-c34af1432a4f/2d0a93a4-59a8-4788-93ce-c34af1432a4f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.349818] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5894129b-7fcb-46f4-ad92-d19970e92104 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.365956] env[65107]: DEBUG nova.compute.manager [req-6ed76d86-3807-4f94-9a1d-296c07dad23e req-93ab6df1-99fa-4c94-845a-04d969cd22a1 service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Received event network-vif-plugged-860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 835.366304] env[65107]: DEBUG oslo_concurrency.lockutils [req-6ed76d86-3807-4f94-9a1d-296c07dad23e req-93ab6df1-99fa-4c94-845a-04d969cd22a1 service nova] Acquiring lock "40474981-eeef-492e-8a8b-aaea7f554c02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.366423] env[65107]: DEBUG oslo_concurrency.lockutils [req-6ed76d86-3807-4f94-9a1d-296c07dad23e req-93ab6df1-99fa-4c94-845a-04d969cd22a1 service nova] Lock "40474981-eeef-492e-8a8b-aaea7f554c02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.366587] env[65107]: DEBUG oslo_concurrency.lockutils [req-6ed76d86-3807-4f94-9a1d-296c07dad23e req-93ab6df1-99fa-4c94-845a-04d969cd22a1 service nova] Lock "40474981-eeef-492e-8a8b-aaea7f554c02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.366752] env[65107]: DEBUG nova.compute.manager [req-6ed76d86-3807-4f94-9a1d-296c07dad23e req-93ab6df1-99fa-4c94-845a-04d969cd22a1 service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] No waiting events found dispatching network-vif-plugged-860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 835.366967] env[65107]: WARNING nova.compute.manager [req-6ed76d86-3807-4f94-9a1d-296c07dad23e req-93ab6df1-99fa-4c94-845a-04d969cd22a1 service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Received unexpected event network-vif-plugged-860d5e06-e063-471a-8b43-cda8c6bd94b8 for instance with vm_state building and task_state spawning. [ 835.368570] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.368954] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.382247] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 835.382247] env[65107]: value = "task-5102825" [ 835.382247] env[65107]: _type = "Task" [ 835.382247] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.388669] env[65107]: DEBUG nova.compute.manager [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 835.388669] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.389704] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe1d8cb-7282-44cf-967f-a14930b881a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.397329] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.407124] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.407124] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fdc9ca7-2514-4d05-9ab3-8957bd6f0af4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.415036] env[65107]: DEBUG oslo_vmware.api [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 835.415036] env[65107]: value = "task-5102826" [ 835.415036] env[65107]: _type = "Task" [ 835.415036] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.425310] env[65107]: DEBUG oslo_vmware.api [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102826, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.437355] env[65107]: WARNING neutronclient.v2_0.client [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.438062] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.438806] env[65107]: WARNING openstack [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.542899] env[65107]: DEBUG nova.network.neutron [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [{"id": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "address": "fa:16:3e:98:5d:68", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb44410c9-c2", "ovs_interfaceid": "b44410c9-c28d-4e13-879d-3e495fde9ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 835.664627] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.664796] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 835.665109] env[65107]: DEBUG nova.network.neutron [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 835.684250] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102824, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.750405] env[65107]: DEBUG nova.compute.utils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 835.751944] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 835.752212] env[65107]: DEBUG nova.network.neutron [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 835.752481] env[65107]: WARNING neutronclient.v2_0.client [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.753109] env[65107]: WARNING neutronclient.v2_0.client [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.753422] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.753703] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.762494] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "780f910f-4f73-41ba-a795-0daae3097314" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.762734] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "780f910f-4f73-41ba-a795-0daae3097314" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.762929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "780f910f-4f73-41ba-a795-0daae3097314-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.763125] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "780f910f-4f73-41ba-a795-0daae3097314-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.763294] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "780f910f-4f73-41ba-a795-0daae3097314-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.767677] env[65107]: INFO nova.compute.manager [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Terminating instance [ 835.803903] env[65107]: DEBUG nova.policy [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '513e39583cc4477c828531bb3250de31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f883a2e25da34a548b020184acb9245e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 835.808553] env[65107]: INFO nova.compute.manager [-] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Took 1.32 seconds to deallocate network for instance. [ 835.893410] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102825, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.925955] env[65107]: DEBUG oslo_vmware.api [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102826, 'name': PowerOffVM_Task, 'duration_secs': 0.301032} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.926289] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.926479] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.926714] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-148eac8b-617f-4a59-ae0c-aa27d86b6339 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.001555] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 836.001795] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 836.001977] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleting the datastore file [datastore1] d2d0e18c-4839-416c-b363-b93996e8ba7f {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.002319] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e325dfa-65a8-43bb-a1dc-de3c58e0ff90 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.009920] env[65107]: DEBUG oslo_vmware.api [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 836.009920] env[65107]: value = "task-5102828" [ 836.009920] env[65107]: _type = "Task" [ 836.009920] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.021291] env[65107]: DEBUG oslo_vmware.api [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.046373] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-d2d0e18c-4839-416c-b363-b93996e8ba7f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 836.060459] env[65107]: DEBUG nova.compute.manager [req-58a7c1db-4beb-4f58-a7b7-6df54ab48cc7 req-75b4b7cf-3b02-4d3f-a4a7-caa246ef363e service nova] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Received event network-vif-deleted-487dbf21-efa0-47ad-8919-f245c183a8ab {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 836.168432] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.168964] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.179376] env[65107]: DEBUG nova.network.neutron [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Successfully created port: c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 836.193193] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102824, 'name': ReconfigVM_Task, 'duration_secs': 0.564348} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.193589] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 805e2210-fb58-48da-be63-5d548b2f8a2f/805e2210-fb58-48da-be63-5d548b2f8a2f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.194347] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f656b437-9186-41e1-82f5-2fa15ae059e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.201435] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 836.201435] env[65107]: value = "task-5102829" [ 836.201435] env[65107]: _type = "Task" [ 836.201435] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.211123] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102829, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.230523] env[65107]: DEBUG nova.network.neutron [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 836.259499] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.259911] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.266822] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 836.273848] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "refresh_cache-780f910f-4f73-41ba-a795-0daae3097314" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.274184] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquired lock "refresh_cache-780f910f-4f73-41ba-a795-0daae3097314" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.274436] env[65107]: DEBUG nova.network.neutron [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 836.316283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.385014] env[65107]: WARNING neutronclient.v2_0.client [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.385248] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.385582] env[65107]: WARNING openstack [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.409825] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102825, 'name': ReconfigVM_Task, 'duration_secs': 0.523711} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.410402] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 2d0a93a4-59a8-4788-93ce-c34af1432a4f/2d0a93a4-59a8-4788-93ce-c34af1432a4f.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.411269] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b16d3811-ef6c-404d-864f-44d95b65630a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.420513] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 836.420513] env[65107]: value = "task-5102830" [ 836.420513] env[65107]: _type = "Task" [ 836.420513] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.439334] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102830, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.528491] env[65107]: DEBUG oslo_vmware.api [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5102828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164169} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.528784] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.528986] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.531488] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.531488] env[65107]: INFO nova.compute.manager [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 836.531593] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 836.531786] env[65107]: DEBUG nova.compute.manager [-] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 836.532316] env[65107]: DEBUG nova.network.neutron [-] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 836.533073] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.533225] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.535409] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.550570] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bdab4f60-f47c-44a6-95af-248a830a47c8 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-d2d0e18c-4839-416c-b363-b93996e8ba7f-97579a1b-69c4-4b0c-b832-426f4d60592d" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.832s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.572603] env[65107]: DEBUG nova.network.neutron [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Updating instance_info_cache with network_info: [{"id": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "address": "fa:16:3e:8e:07:f3", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860d5e06-e0", "ovs_interfaceid": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.628099] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.716226] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102829, 'name': Rename_Task, 'duration_secs': 0.204791} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.716528] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.716782] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fda9aabf-17e7-4b60-be1d-2031a80b6b78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.724828] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 836.724828] env[65107]: value = "task-5102831" [ 836.724828] env[65107]: _type = "Task" [ 836.724828] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.738052] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.777674] env[65107]: WARNING neutronclient.v2_0.client [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.778385] env[65107]: WARNING openstack [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.778772] env[65107]: WARNING openstack [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.848164] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b546e16-497c-4a48-b2c8-2d1dc1f616d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.859166] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5205691f-4c87-4df6-acc3-9e7a8ada6f05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.894782] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187a9ea5-1bdc-4744-8bb5-2485dbca5032 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.903438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c536f668-26bc-4263-8807-55dce9b43199 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.918157] env[65107]: DEBUG nova.compute.provider_tree [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.934416] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102830, 'name': Rename_Task, 'duration_secs': 0.302233} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.934770] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.935041] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2847c02f-d658-41a8-8f20-cbed78702aed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.943158] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 836.943158] env[65107]: value = "task-5102832" [ 836.943158] env[65107]: _type = "Task" [ 836.943158] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.954567] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.076189] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.076189] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Instance network_info: |[{"id": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "address": "fa:16:3e:8e:07:f3", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860d5e06-e0", "ovs_interfaceid": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 837.076712] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:07:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f49a7d-c6e5-404f-b71a-91d8c070cd18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '860d5e06-e063-471a-8b43-cda8c6bd94b8', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.084964] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Creating folder: Project (f883a2e25da34a548b020184acb9245e). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.084964] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed69a1f7-8c63-40e6-9ebe-d592cc8eeb63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.096356] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Created folder: Project (f883a2e25da34a548b020184acb9245e) in parent group-v992574. [ 837.096563] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Creating folder: Instances. Parent ref: group-v992744. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.096808] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18d22d35-5f65-46da-8046-659f68c0365c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.109063] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Created folder: Instances in parent group-v992744. [ 837.109386] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.109598] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.110560] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c590009b-23d5-41c1-ae4f-e84637c01955 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.135944] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.135944] env[65107]: value = "task-5102835" [ 837.135944] env[65107]: _type = "Task" [ 837.135944] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.145085] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102835, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.146162] env[65107]: DEBUG nova.network.neutron [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 837.213939] env[65107]: DEBUG nova.network.neutron [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.236031] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102831, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.276510] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 837.424055] env[65107]: DEBUG nova.scheduler.client.report [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 837.443568] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 837.443568] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 837.443568] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 837.444462] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 837.444462] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 837.444462] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 837.444462] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.444462] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 837.444765] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 837.444765] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 837.444850] env[65107]: DEBUG nova.virt.hardware [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 837.445849] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c923dd-6de8-47f9-9bb6-7c6ddddfe1b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.464213] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102832, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.465496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c330181-07a9-4967-857d-493ba04ebc99 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.601069] env[65107]: DEBUG nova.network.neutron [-] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.656214] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102835, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.665536] env[65107]: DEBUG nova.compute.manager [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Received event network-changed-860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 837.665840] env[65107]: DEBUG nova.compute.manager [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Refreshing instance network info cache due to event network-changed-860d5e06-e063-471a-8b43-cda8c6bd94b8. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 837.666395] env[65107]: DEBUG oslo_concurrency.lockutils [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Acquiring lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.666625] env[65107]: DEBUG oslo_concurrency.lockutils [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Acquired lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.668097] env[65107]: DEBUG nova.network.neutron [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Refreshing network info cache for port 860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 837.717753] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Releasing lock "refresh_cache-780f910f-4f73-41ba-a795-0daae3097314" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.718210] env[65107]: DEBUG nova.compute.manager [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 837.718493] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 837.719698] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1951339-e7d5-4362-8c4b-779a2d4e00c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.740836] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 837.745121] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95ed500a-ab0e-4a0e-82c6-83d401f2c633 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.748678] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102831, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.755960] env[65107]: DEBUG oslo_vmware.api [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 837.755960] env[65107]: value = "task-5102836" [ 837.755960] env[65107]: _type = "Task" [ 837.755960] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.768695] env[65107]: DEBUG oslo_vmware.api [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.823702] env[65107]: DEBUG nova.network.neutron [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Successfully updated port: c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 837.930777] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.689s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.931194] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 837.935323] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 39.599s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.935641] env[65107]: DEBUG nova.objects.instance [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 837.956517] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102832, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.094106] env[65107]: DEBUG nova.compute.manager [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Received event network-vif-plugged-c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 838.094279] env[65107]: DEBUG oslo_concurrency.lockutils [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Acquiring lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.094561] env[65107]: DEBUG oslo_concurrency.lockutils [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.094647] env[65107]: DEBUG oslo_concurrency.lockutils [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.094813] env[65107]: DEBUG nova.compute.manager [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] No waiting events found dispatching network-vif-plugged-c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 838.094974] env[65107]: WARNING nova.compute.manager [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Received unexpected event network-vif-plugged-c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 for instance with vm_state building and task_state spawning. [ 838.095228] env[65107]: DEBUG nova.compute.manager [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Received event network-changed-c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 838.095459] env[65107]: DEBUG nova.compute.manager [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Refreshing instance network info cache due to event network-changed-c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 838.095613] env[65107]: DEBUG oslo_concurrency.lockutils [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Acquiring lock "refresh_cache-21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.095708] env[65107]: DEBUG oslo_concurrency.lockutils [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Acquired lock "refresh_cache-21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.096281] env[65107]: DEBUG nova.network.neutron [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Refreshing network info cache for port c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 838.104484] env[65107]: INFO nova.compute.manager [-] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Took 1.57 seconds to deallocate network for instance. [ 838.150016] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102835, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.171307] env[65107]: WARNING neutronclient.v2_0.client [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.172406] env[65107]: WARNING openstack [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.172554] env[65107]: WARNING openstack [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.241232] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102831, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.267895] env[65107]: DEBUG oslo_vmware.api [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102836, 'name': PowerOffVM_Task, 'duration_secs': 0.167575} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.268300] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 838.268559] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 838.268959] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72b4775c-15a1-429e-8937-4bf760b670bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.306154] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 838.306154] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 838.306585] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Deleting the datastore file [datastore2] 780f910f-4f73-41ba-a795-0daae3097314 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 838.306844] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e32560d8-6f93-45de-b443-9023e3e86e80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.317011] env[65107]: WARNING openstack [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.317609] env[65107]: WARNING openstack [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.325824] env[65107]: DEBUG oslo_vmware.api [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for the task: (returnval){ [ 838.325824] env[65107]: value = "task-5102838" [ 838.325824] env[65107]: _type = "Task" [ 838.325824] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.327483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "refresh_cache-21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.338449] env[65107]: DEBUG oslo_vmware.api [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.403264] env[65107]: WARNING neutronclient.v2_0.client [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.404101] env[65107]: WARNING openstack [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.404692] env[65107]: WARNING openstack [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.442197] env[65107]: DEBUG nova.compute.utils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 838.446853] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 838.447152] env[65107]: DEBUG nova.network.neutron [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 838.447522] env[65107]: WARNING neutronclient.v2_0.client [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.447851] env[65107]: WARNING neutronclient.v2_0.client [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.448432] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.448828] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.473030] env[65107]: DEBUG oslo_vmware.api [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102832, 'name': PowerOnVM_Task, 'duration_secs': 1.475516} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.473367] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.473598] env[65107]: INFO nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Took 8.81 seconds to spawn the instance on the hypervisor. [ 838.473809] env[65107]: DEBUG nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 838.474934] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c005a86a-b444-47eb-b977-583dec2898a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.503927] env[65107]: DEBUG nova.policy [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '513e39583cc4477c828531bb3250de31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f883a2e25da34a548b020184acb9245e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 838.514053] env[65107]: DEBUG nova.network.neutron [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Updated VIF entry in instance network info cache for port 860d5e06-e063-471a-8b43-cda8c6bd94b8. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 838.514432] env[65107]: DEBUG nova.network.neutron [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Updating instance_info_cache with network_info: [{"id": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "address": "fa:16:3e:8e:07:f3", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860d5e06-e0", "ovs_interfaceid": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 838.599344] env[65107]: WARNING neutronclient.v2_0.client [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.600148] env[65107]: WARNING openstack [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.600492] env[65107]: WARNING openstack [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.611080] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.638850] env[65107]: DEBUG nova.network.neutron [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 838.651401] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102835, 'name': CreateVM_Task, 'duration_secs': 1.478736} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.651619] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.652160] env[65107]: WARNING neutronclient.v2_0.client [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.652559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.652801] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.653068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 838.653305] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90a296d7-644b-44e6-a13a-a7cc5e8123fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.660057] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 838.660057] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b84c5a-f809-a35b-5b02-24c112d2277a" [ 838.660057] env[65107]: _type = "Task" [ 838.660057] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.670697] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b84c5a-f809-a35b-5b02-24c112d2277a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.740137] env[65107]: DEBUG oslo_vmware.api [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102831, 'name': PowerOnVM_Task, 'duration_secs': 1.553078} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.740137] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.740277] env[65107]: INFO nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Took 11.25 seconds to spawn the instance on the hypervisor. [ 838.740392] env[65107]: DEBUG nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 838.741247] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fcda36-0b96-4de8-85b1-57108bb45527 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.798842] env[65107]: DEBUG nova.network.neutron [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 838.833116] env[65107]: DEBUG nova.network.neutron [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Successfully created port: 0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 838.839829] env[65107]: DEBUG oslo_vmware.api [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Task: {'id': task-5102838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120975} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.839829] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.840094] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 838.840169] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 838.840388] env[65107]: INFO nova.compute.manager [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Took 1.12 seconds to destroy the instance on the hypervisor. [ 838.840747] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 838.840990] env[65107]: DEBUG nova.compute.manager [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 838.841133] env[65107]: DEBUG nova.network.neutron [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 838.841393] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.841949] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.842231] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.865390] env[65107]: DEBUG nova.network.neutron [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 838.865614] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.948032] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 838.958721] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3c467b71-e73b-4905-869e-eb7c13ce622d tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.959840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.418s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.961453] env[65107]: INFO nova.compute.claims [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.000017] env[65107]: INFO nova.compute.manager [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Took 43.67 seconds to build instance. [ 839.018409] env[65107]: DEBUG oslo_concurrency.lockutils [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] Releasing lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.019797] env[65107]: DEBUG nova.compute.manager [req-69cd5921-90c7-4a6f-a5a7-e256698889e0 req-ba540456-f544-4922-bbef-b5617ccc66ba service nova] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Received event network-vif-deleted-b44410c9-c28d-4e13-879d-3e495fde9ff4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 839.172361] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b84c5a-f809-a35b-5b02-24c112d2277a, 'name': SearchDatastore_Task, 'duration_secs': 0.010921} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.173136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.173401] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.173734] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.173943] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.174157] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.174436] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b9008f4-d42a-4e58-99a6-b8eca0c19d1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.186039] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.186312] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.187236] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af6069af-99b1-41ae-ab6a-d759f77ec177 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.193664] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 839.193664] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52209237-c1aa-3910-50fe-a890b73d10fc" [ 839.193664] env[65107]: _type = "Task" [ 839.193664] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.203477] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52209237-c1aa-3910-50fe-a890b73d10fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.264211] env[65107]: INFO nova.compute.manager [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Took 47.74 seconds to build instance. [ 839.302026] env[65107]: DEBUG oslo_concurrency.lockutils [req-a148717a-a454-4e67-bf5b-156583bb3c6f req-6fbb781e-514c-401b-aa98-8969d3d9fc5d service nova] Releasing lock "refresh_cache-21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.302924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "refresh_cache-21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.302924] env[65107]: DEBUG nova.network.neutron [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 839.367949] env[65107]: DEBUG nova.network.neutron [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.500829] env[65107]: DEBUG oslo_concurrency.lockutils [None req-604a1359-c15e-4ae8-b790-ce5a93a92d76 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.222s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.705616] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52209237-c1aa-3910-50fe-a890b73d10fc, 'name': SearchDatastore_Task, 'duration_secs': 0.010025} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.706586] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4118c14f-5773-4879-8d5e-1e9aa3eb2137 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.712766] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 839.712766] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e34f99-87df-96c5-99c2-f0a0e5384c2d" [ 839.712766] env[65107]: _type = "Task" [ 839.712766] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.721221] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e34f99-87df-96c5-99c2-f0a0e5384c2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.765038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-36649a80-8cf7-45ac-9fa8-1f31b031d378 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.258s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.804049] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc78070-8600-4449-9e05-900fd40dc49f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.809077] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.809397] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.821672] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Suspending the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 839.822365] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-de44f3ae-0928-4f61-9e77-8c8e0c3d62f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.831094] env[65107]: DEBUG oslo_vmware.api [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 839.831094] env[65107]: value = "task-5102839" [ 839.831094] env[65107]: _type = "Task" [ 839.831094] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.844043] env[65107]: DEBUG oslo_vmware.api [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102839, 'name': SuspendVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.856818] env[65107]: DEBUG nova.network.neutron [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 839.871775] env[65107]: INFO nova.compute.manager [-] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Took 1.03 seconds to deallocate network for instance. [ 839.877973] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.878376] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.960677] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 839.992257] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 839.992582] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 839.992759] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 839.993079] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 839.993395] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 839.993477] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 839.993720] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.993911] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 839.994117] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 839.994285] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 839.994485] env[65107]: DEBUG nova.virt.hardware [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 839.996082] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11363566-c6c8-48cc-b577-679d4072dc17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.004310] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 840.009481] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fe77dd-35a2-415a-a4ad-620013e5124c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.067035] env[65107]: WARNING neutronclient.v2_0.client [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.068149] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.068149] env[65107]: WARNING openstack [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.226584] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e34f99-87df-96c5-99c2-f0a0e5384c2d, 'name': SearchDatastore_Task, 'duration_secs': 0.029104} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.226843] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 840.227124] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 40474981-eeef-492e-8a8b-aaea7f554c02/40474981-eeef-492e-8a8b-aaea7f554c02.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.227491] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13df27f7-9a8b-4c68-8de6-f9d31207572f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.237189] env[65107]: DEBUG nova.network.neutron [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Updating instance_info_cache with network_info: [{"id": "c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7", "address": "fa:16:3e:b3:aa:1e", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9e1f243-9b", "ovs_interfaceid": "c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 840.241026] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 840.241026] env[65107]: value = "task-5102840" [ 840.241026] env[65107]: _type = "Task" [ 840.241026] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.251032] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.270485] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 840.348306] env[65107]: DEBUG oslo_vmware.api [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102839, 'name': SuspendVM_Task} progress is 62%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.381768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.391618] env[65107]: DEBUG nova.compute.manager [req-f2f91b46-72e8-4659-bc85-a94786a12a07 req-ecf4a2b6-d52c-4648-b8bc-6194df2d9f8e service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Received event network-vif-plugged-0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 840.391861] env[65107]: DEBUG oslo_concurrency.lockutils [req-f2f91b46-72e8-4659-bc85-a94786a12a07 req-ecf4a2b6-d52c-4648-b8bc-6194df2d9f8e service nova] Acquiring lock "9a46ff51-4035-4709-ae6c-0fb5397200ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.392151] env[65107]: DEBUG oslo_concurrency.lockutils [req-f2f91b46-72e8-4659-bc85-a94786a12a07 req-ecf4a2b6-d52c-4648-b8bc-6194df2d9f8e service nova] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 840.392290] env[65107]: DEBUG oslo_concurrency.lockutils [req-f2f91b46-72e8-4659-bc85-a94786a12a07 req-ecf4a2b6-d52c-4648-b8bc-6194df2d9f8e service nova] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.392451] env[65107]: DEBUG nova.compute.manager [req-f2f91b46-72e8-4659-bc85-a94786a12a07 req-ecf4a2b6-d52c-4648-b8bc-6194df2d9f8e service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] No waiting events found dispatching network-vif-plugged-0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 840.392710] env[65107]: WARNING nova.compute.manager [req-f2f91b46-72e8-4659-bc85-a94786a12a07 req-ecf4a2b6-d52c-4648-b8bc-6194df2d9f8e service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Received unexpected event network-vif-plugged-0a1d113a-256a-4a98-a5c6-45d4fe610599 for instance with vm_state building and task_state spawning. [ 840.446058] env[65107]: DEBUG nova.network.neutron [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Successfully updated port: 0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 840.534841] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.588209] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac5852b-7f83-433c-81db-bd629d47c865 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.597682] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0b21e6-81f0-460c-8b2c-f84f4ce4ce06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.630353] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d063c99-c16c-4136-9203-e0ef1afc1303 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.643088] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f482114e-cd4c-45c9-9735-0969bfb82dc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.665149] env[65107]: DEBUG nova.compute.provider_tree [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.741945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "refresh_cache-21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 840.742461] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Instance network_info: |[{"id": "c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7", "address": "fa:16:3e:b3:aa:1e", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9e1f243-9b", "ovs_interfaceid": "c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 840.743460] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:aa:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f49a7d-c6e5-404f-b71a-91d8c070cd18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.751585] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 840.756498] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.756928] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22e19df7-fc6e-4671-a7ba-d92f2b51777a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.782562] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102840, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.790521] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.790521] env[65107]: value = "task-5102841" [ 840.790521] env[65107]: _type = "Task" [ 840.790521] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.801058] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102841, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.802041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.844315] env[65107]: DEBUG oslo_vmware.api [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102839, 'name': SuspendVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.950856] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "refresh_cache-9a46ff51-4035-4709-ae6c-0fb5397200ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.951202] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "refresh_cache-9a46ff51-4035-4709-ae6c-0fb5397200ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 840.951611] env[65107]: DEBUG nova.network.neutron [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 841.167935] env[65107]: DEBUG nova.scheduler.client.report [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 841.254218] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.791774} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.254511] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 40474981-eeef-492e-8a8b-aaea7f554c02/40474981-eeef-492e-8a8b-aaea7f554c02.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.254805] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.255392] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ffba3ea3-c4e8-4d5b-a086-c28b4c40f8db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.263087] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 841.263087] env[65107]: value = "task-5102842" [ 841.263087] env[65107]: _type = "Task" [ 841.263087] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.276245] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.302339] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102841, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.342557] env[65107]: DEBUG oslo_vmware.api [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102839, 'name': SuspendVM_Task, 'duration_secs': 1.132475} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.342833] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Suspended the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 841.343023] env[65107]: DEBUG nova.compute.manager [None req-ccb01f12-b7ef-4bd9-842e-e0ab099a2b96 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 841.343786] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5593f61f-1b98-4dbb-aad0-f984c6d03e43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.454948] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.455365] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.504378] env[65107]: DEBUG nova.network.neutron [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 841.526723] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.528514] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.597182] env[65107]: WARNING neutronclient.v2_0.client [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 841.597989] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.598548] env[65107]: WARNING openstack [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.673615] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.713s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.673615] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 841.676245] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.973s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.676493] env[65107]: DEBUG nova.objects.instance [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lazy-loading 'resources' on Instance uuid 1253473b-d4bc-49ac-9724-9aa1e7a8f038 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.693147] env[65107]: DEBUG nova.network.neutron [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Updating instance_info_cache with network_info: [{"id": "0a1d113a-256a-4a98-a5c6-45d4fe610599", "address": "fa:16:3e:52:d6:fa", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1d113a-25", "ovs_interfaceid": "0a1d113a-256a-4a98-a5c6-45d4fe610599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.773109] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081991} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.773422] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.774292] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237cece3-c133-49e9-b770-fbbe1258b217 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.798440] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 40474981-eeef-492e-8a8b-aaea7f554c02/40474981-eeef-492e-8a8b-aaea7f554c02.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.798831] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c58d01a1-af81-44a4-90c8-c7a2dce9d03b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.822715] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102841, 'name': CreateVM_Task, 'duration_secs': 0.546109} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.824099] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 841.824493] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 841.824493] env[65107]: value = "task-5102843" [ 841.824493] env[65107]: _type = "Task" [ 841.824493] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.824952] env[65107]: WARNING neutronclient.v2_0.client [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 841.825331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.825488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 841.825812] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 841.826135] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23091f35-e20e-4227-96ba-e03135e34c73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.837946] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.838329] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 841.838329] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f15031-7d6d-858f-3529-6bbb1737b6f8" [ 841.838329] env[65107]: _type = "Task" [ 841.838329] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.848157] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f15031-7d6d-858f-3529-6bbb1737b6f8, 'name': SearchDatastore_Task, 'duration_secs': 0.011285} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.848479] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 841.848712] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.848944] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.849137] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 841.849391] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.849702] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a7305ff-bd27-4269-84a9-0d9f507d2f72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.859048] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.859372] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.860140] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12071b89-3c12-4085-96bc-585b48456644 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.866555] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 841.866555] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52833f9a-c7f5-1e71-a9e8-61f9911e5065" [ 841.866555] env[65107]: _type = "Task" [ 841.866555] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.875586] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52833f9a-c7f5-1e71-a9e8-61f9911e5065, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.062708] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "805e2210-fb58-48da-be63-5d548b2f8a2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.062868] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.063185] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "805e2210-fb58-48da-be63-5d548b2f8a2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.063432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.063649] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.066293] env[65107]: INFO nova.compute.manager [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Terminating instance [ 842.182911] env[65107]: DEBUG nova.compute.utils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 842.186940] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 842.187218] env[65107]: DEBUG nova.network.neutron [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 842.187546] env[65107]: WARNING neutronclient.v2_0.client [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.187842] env[65107]: WARNING neutronclient.v2_0.client [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.188501] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.188749] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.196048] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "refresh_cache-9a46ff51-4035-4709-ae6c-0fb5397200ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 842.196391] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Instance network_info: |[{"id": "0a1d113a-256a-4a98-a5c6-45d4fe610599", "address": "fa:16:3e:52:d6:fa", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1d113a-25", "ovs_interfaceid": "0a1d113a-256a-4a98-a5c6-45d4fe610599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 842.197481] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:d6:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f49a7d-c6e5-404f-b71a-91d8c070cd18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a1d113a-256a-4a98-a5c6-45d4fe610599', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.206404] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 842.207064] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.207328] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9f03b75-3a4d-4ec8-a248-837d29553655 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.233040] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.233040] env[65107]: value = "task-5102844" [ 842.233040] env[65107]: _type = "Task" [ 842.233040] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.243728] env[65107]: DEBUG nova.policy [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ccab6477dd44878f78a9d6428d3cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4778fe9152224fd29d1f6220a19b5a36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 842.249880] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102844, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.337950] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102843, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.385529] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52833f9a-c7f5-1e71-a9e8-61f9911e5065, 'name': SearchDatastore_Task, 'duration_secs': 0.010596} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.391724] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afd8babd-7049-483d-8eff-efde7e43b4fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.399164] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 842.399164] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522a2656-8ac3-4d34-527e-f9881665179a" [ 842.399164] env[65107]: _type = "Task" [ 842.399164] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.409149] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522a2656-8ac3-4d34-527e-f9881665179a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.572373] env[65107]: DEBUG nova.compute.manager [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 842.572373] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.572745] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b34b53-f6df-4daf-9795-067097079735 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.582974] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.582974] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d41f241a-c7fb-4e18-9d94-e7f9580c4d5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.588742] env[65107]: DEBUG nova.network.neutron [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Successfully created port: 8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 842.597476] env[65107]: DEBUG oslo_vmware.api [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 842.597476] env[65107]: value = "task-5102845" [ 842.597476] env[65107]: _type = "Task" [ 842.597476] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.613755] env[65107]: DEBUG oslo_vmware.api [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.697048] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 842.743506] env[65107]: DEBUG nova.compute.manager [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Received event network-changed-0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 842.743737] env[65107]: DEBUG nova.compute.manager [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Refreshing instance network info cache due to event network-changed-0a1d113a-256a-4a98-a5c6-45d4fe610599. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 842.743955] env[65107]: DEBUG oslo_concurrency.lockutils [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Acquiring lock "refresh_cache-9a46ff51-4035-4709-ae6c-0fb5397200ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.744691] env[65107]: DEBUG oslo_concurrency.lockutils [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Acquired lock "refresh_cache-9a46ff51-4035-4709-ae6c-0fb5397200ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.744691] env[65107]: DEBUG nova.network.neutron [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Refreshing network info cache for port 0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 842.753889] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102844, 'name': CreateVM_Task, 'duration_secs': 0.401986} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.754043] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.754583] env[65107]: WARNING neutronclient.v2_0.client [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.754939] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.755315] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.755402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 842.755694] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13c7cd22-8f14-44ba-87c9-6f5d425f0c47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.765082] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 842.765082] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523d445a-5cce-32be-91c2-965804d4fdad" [ 842.765082] env[65107]: _type = "Task" [ 842.765082] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.777511] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523d445a-5cce-32be-91c2-965804d4fdad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.780231] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636203f2-a71c-4a68-88c3-32c80f556925 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.788879] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbedfcf-da79-4ca7-b682-498c08f9c315 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.827352] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53547c59-653e-4e4e-8ea2-7cee1e46e24a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.842608] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e50404b-1c05-48fa-9cb2-5be2adc9bc1e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.846823] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102843, 'name': ReconfigVM_Task, 'duration_secs': 0.694745} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.847171] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 40474981-eeef-492e-8a8b-aaea7f554c02/40474981-eeef-492e-8a8b-aaea7f554c02.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.848260] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6827642-bc2f-497e-911e-37159677f615 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.859340] env[65107]: DEBUG nova.compute.provider_tree [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.865024] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 842.865024] env[65107]: value = "task-5102846" [ 842.865024] env[65107]: _type = "Task" [ 842.865024] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.874606] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102846, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.911245] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522a2656-8ac3-4d34-527e-f9881665179a, 'name': SearchDatastore_Task, 'duration_secs': 0.01287} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.911830] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 842.912148] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c/21c471cd-2435-4e4e-82cc-8cdf03f1fa6c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.912462] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26901412-9a4a-4d65-9982-af8b7a47cdef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.920172] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 842.920172] env[65107]: value = "task-5102847" [ 842.920172] env[65107]: _type = "Task" [ 842.920172] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.931781] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.108862] env[65107]: DEBUG oslo_vmware.api [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102845, 'name': PowerOffVM_Task, 'duration_secs': 0.217027} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.109342] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.109452] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.109913] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54fd88e9-1a6c-4afd-ab30-1a1773d25447 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.184778] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.185035] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.185207] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Deleting the datastore file [datastore2] 805e2210-fb58-48da-be63-5d548b2f8a2f {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.185605] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ce046fe-9fb9-470b-8762-4743ea3782a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.195884] env[65107]: DEBUG oslo_vmware.api [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for the task: (returnval){ [ 843.195884] env[65107]: value = "task-5102849" [ 843.195884] env[65107]: _type = "Task" [ 843.195884] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.213077] env[65107]: DEBUG oslo_vmware.api [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.247705] env[65107]: WARNING neutronclient.v2_0.client [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.248097] env[65107]: WARNING openstack [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.248505] env[65107]: WARNING openstack [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.283521] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523d445a-5cce-32be-91c2-965804d4fdad, 'name': SearchDatastore_Task, 'duration_secs': 0.011779} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.284425] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 843.284425] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.284425] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.284730] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.284814] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.285186] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45052d21-f5bb-494a-acda-fe4baadcb8ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.314091] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.314429] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.315649] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80820b86-4f8a-4661-b419-d43e603ee911 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.324178] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 843.324178] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c98207-e1a0-e1d3-1d48-067ed13e45b6" [ 843.324178] env[65107]: _type = "Task" [ 843.324178] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.332669] env[65107]: DEBUG nova.compute.manager [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 843.338186] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd32d36c-1af1-4c81-8c0a-3a825c332e31 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.348787] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c98207-e1a0-e1d3-1d48-067ed13e45b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.382484] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102846, 'name': Rename_Task, 'duration_secs': 0.179827} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.383141] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 843.383141] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6b8ee6d-b1b0-455b-824c-809667eb886c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.393808] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 843.393808] env[65107]: value = "task-5102850" [ 843.393808] env[65107]: _type = "Task" [ 843.393808] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.400296] env[65107]: WARNING openstack [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.400983] env[65107]: WARNING openstack [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.413024] env[65107]: ERROR nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] [req-76a3aafe-3010-49d9-85be-6101c390e5b0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-76a3aafe-3010-49d9-85be-6101c390e5b0"}]} [ 843.426096] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102850, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.436044] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50908} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.436044] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c/21c471cd-2435-4e4e-82cc-8cdf03f1fa6c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.436044] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.436044] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f10fb45-9860-4ff8-860c-3abc97a20c18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.443982] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 843.443982] env[65107]: value = "task-5102851" [ 843.443982] env[65107]: _type = "Task" [ 843.443982] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.453385] env[65107]: DEBUG nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 843.461883] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.470318] env[65107]: DEBUG nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 843.470606] env[65107]: DEBUG nova.compute.provider_tree [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.484824] env[65107]: DEBUG nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 843.513710] env[65107]: WARNING neutronclient.v2_0.client [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.514409] env[65107]: WARNING openstack [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.514783] env[65107]: WARNING openstack [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.525730] env[65107]: DEBUG nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 843.705238] env[65107]: DEBUG nova.network.neutron [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Updated VIF entry in instance network info cache for port 0a1d113a-256a-4a98-a5c6-45d4fe610599. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 843.705604] env[65107]: DEBUG nova.network.neutron [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Updating instance_info_cache with network_info: [{"id": "0a1d113a-256a-4a98-a5c6-45d4fe610599", "address": "fa:16:3e:52:d6:fa", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a1d113a-25", "ovs_interfaceid": "0a1d113a-256a-4a98-a5c6-45d4fe610599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 843.713685] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 843.715844] env[65107]: DEBUG oslo_vmware.api [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Task: {'id': task-5102849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335327} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.716394] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.716627] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.716745] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.716907] env[65107]: INFO nova.compute.manager [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 843.717175] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 843.719923] env[65107]: DEBUG nova.compute.manager [-] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 843.720032] env[65107]: DEBUG nova.network.neutron [-] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 843.720267] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.720777] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.721104] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.742507] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52723b0e-d7aa-b1fb-edde-3f9783ca3533/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 843.744293] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbad6de1-4e22-40de-89ae-4bbcc613a634 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.751804] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52723b0e-d7aa-b1fb-edde-3f9783ca3533/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 843.752035] env[65107]: ERROR oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52723b0e-d7aa-b1fb-edde-3f9783ca3533/disk-0.vmdk due to incomplete transfer. [ 843.754739] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e2593eb2-a9c4-4c6f-8273-a8e56a9f7e89 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.758667] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 843.758941] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 843.759154] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 843.759377] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 843.759567] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 843.759763] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 843.759942] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.760115] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 843.760277] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 843.760439] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 843.760633] env[65107]: DEBUG nova.virt.hardware [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 843.761824] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a0d690-a5f4-4147-9312-6f9d6ec2e467 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.770900] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57061476-0d35-455f-87db-7c0395239a84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.775659] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.780416] env[65107]: DEBUG oslo_vmware.rw_handles [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52723b0e-d7aa-b1fb-edde-3f9783ca3533/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 843.780722] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Uploaded image 9b846d61-6f7a-470d-bef4-76e583435ca3 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 843.783110] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 843.784269] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-55342203-e0ab-46bc-8147-0481d8ae486a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.803608] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 843.803608] env[65107]: value = "task-5102852" [ 843.803608] env[65107]: _type = "Task" [ 843.803608] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.812369] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102852, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.840222] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c98207-e1a0-e1d3-1d48-067ed13e45b6, 'name': SearchDatastore_Task, 'duration_secs': 0.059378} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.843086] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-703e7a27-a445-48dc-a81a-7f7cfa65a22a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.845352] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquiring lock "7062b119-01a4-45f4-bb52-212178382271" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.846103] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "7062b119-01a4-45f4-bb52-212178382271" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.846206] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquiring lock "7062b119-01a4-45f4-bb52-212178382271-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.846459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "7062b119-01a4-45f4-bb52-212178382271-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.846736] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "7062b119-01a4-45f4-bb52-212178382271-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.856070] env[65107]: INFO nova.compute.manager [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Terminating instance [ 843.856577] env[65107]: INFO nova.compute.manager [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] instance snapshotting [ 843.856882] env[65107]: WARNING nova.compute.manager [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 843.863875] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 843.863875] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522ebe32-06e3-c9d8-9a23-bb5e7ae3722c" [ 843.863875] env[65107]: _type = "Task" [ 843.863875] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.864339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0cef7f-ef56-44ad-a656-59226620bc08 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.882089] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522ebe32-06e3-c9d8-9a23-bb5e7ae3722c, 'name': SearchDatastore_Task, 'duration_secs': 0.010869} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.895733] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 843.896052] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 9a46ff51-4035-4709-ae6c-0fb5397200ab/9a46ff51-4035-4709-ae6c-0fb5397200ab.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 843.899631] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ceb172b-dc38-4104-9582-4cc28ed831ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.905919] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c018fd-882b-4e2d-bf78-7ce1546f1de9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.926281] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102850, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.926281] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 843.926281] env[65107]: value = "task-5102853" [ 843.926281] env[65107]: _type = "Task" [ 843.926281] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.938326] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.954714] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07031} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.954714] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.955213] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05de3b31-7cde-43fb-ad64-f2e6bae252eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.979554] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c/21c471cd-2435-4e4e-82cc-8cdf03f1fa6c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.985200] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21c836bc-4ecb-4051-b7f3-5d69bec98a1e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.009080] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 844.009080] env[65107]: value = "task-5102854" [ 844.009080] env[65107]: _type = "Task" [ 844.009080] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.021046] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102854, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.184843] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deacf22a-71e1-4a06-bba5-54ac203df4dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.197178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a77631-63e8-4e3f-bd4b-2eda586ba86b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.237250] env[65107]: DEBUG oslo_concurrency.lockutils [req-2e06aa57-872c-4e45-97c1-f240d7425da5 req-3404809c-a484-48a6-b335-a4073aa11e63 service nova] Releasing lock "refresh_cache-9a46ff51-4035-4709-ae6c-0fb5397200ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 844.238419] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44e1897-79c3-4109-9b95-68629989b76f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.248265] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb668b3-b2f9-4f30-99f7-3718fb980fbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.266988] env[65107]: DEBUG nova.compute.provider_tree [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.289540] env[65107]: DEBUG nova.network.neutron [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Successfully updated port: 8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 844.321616] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102852, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.364430] env[65107]: DEBUG nova.compute.manager [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 844.364721] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.365068] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25189c05-52b8-4303-9e58-7993bf3d901e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.370590] env[65107]: DEBUG nova.compute.manager [req-e2c718db-bf4b-499a-9555-4deeb518f8bc req-07c74666-ce39-47b5-9b99-4fc059c75f1d service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Received event network-vif-plugged-8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 844.370590] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2c718db-bf4b-499a-9555-4deeb518f8bc req-07c74666-ce39-47b5-9b99-4fc059c75f1d service nova] Acquiring lock "701a1a16-9348-4748-bb24-ee245a76566c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.370977] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2c718db-bf4b-499a-9555-4deeb518f8bc req-07c74666-ce39-47b5-9b99-4fc059c75f1d service nova] Lock "701a1a16-9348-4748-bb24-ee245a76566c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.371195] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2c718db-bf4b-499a-9555-4deeb518f8bc req-07c74666-ce39-47b5-9b99-4fc059c75f1d service nova] Lock "701a1a16-9348-4748-bb24-ee245a76566c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.371495] env[65107]: DEBUG nova.compute.manager [req-e2c718db-bf4b-499a-9555-4deeb518f8bc req-07c74666-ce39-47b5-9b99-4fc059c75f1d service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] No waiting events found dispatching network-vif-plugged-8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 844.371816] env[65107]: WARNING nova.compute.manager [req-e2c718db-bf4b-499a-9555-4deeb518f8bc req-07c74666-ce39-47b5-9b99-4fc059c75f1d service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Received unexpected event network-vif-plugged-8bb68bf5-be11-4086-a291-e8d4eeb2047b for instance with vm_state building and task_state spawning. [ 844.379017] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 844.379017] env[65107]: value = "task-5102855" [ 844.379017] env[65107]: _type = "Task" [ 844.379017] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.392392] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102855, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.411115] env[65107]: DEBUG oslo_vmware.api [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102850, 'name': PowerOnVM_Task, 'duration_secs': 0.570121} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.411515] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.411791] env[65107]: INFO nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Took 9.86 seconds to spawn the instance on the hypervisor. [ 844.412086] env[65107]: DEBUG nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 844.412955] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebfdc70-78b5-490a-8518-9dde85ac2fad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.431034] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 844.431034] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-93c902be-ac0b-4586-b8eb-8cad204cd912 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.443058] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524812} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.445105] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 9a46ff51-4035-4709-ae6c-0fb5397200ab/9a46ff51-4035-4709-ae6c-0fb5397200ab.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 844.445561] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.446914] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 844.446914] env[65107]: value = "task-5102856" [ 844.446914] env[65107]: _type = "Task" [ 844.446914] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.446914] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f068d64-26ec-4252-bf26-a869b0e055a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.459203] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102856, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.461207] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 844.461207] env[65107]: value = "task-5102857" [ 844.461207] env[65107]: _type = "Task" [ 844.461207] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.472260] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.520505] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.536125] env[65107]: DEBUG nova.network.neutron [-] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 844.772031] env[65107]: DEBUG nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 844.796352] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.796352] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 844.796352] env[65107]: DEBUG nova.network.neutron [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 844.819598] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102852, 'name': Destroy_Task, 'duration_secs': 0.580633} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.819904] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroyed the VM [ 844.820171] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 844.820657] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8a49d5be-5324-482e-9286-3d7d33ed5d4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.829239] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 844.829239] env[65107]: value = "task-5102858" [ 844.829239] env[65107]: _type = "Task" [ 844.829239] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.839333] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102858, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.841569] env[65107]: DEBUG nova.compute.manager [req-12ff5fa0-4264-4f14-ab10-22437f2b75d8 req-49faf022-36c3-4b15-9ecc-08269db6a547 service nova] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Received event network-vif-deleted-535ab747-8154-4faf-ae33-4f3a850a5dfb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 844.890073] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102855, 'name': PowerOffVM_Task, 'duration_secs': 0.339508} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.890352] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 844.890574] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 844.890766] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992609', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'name': 'volume-7ab345f6-458d-4bb0-be22-c593a481ce3f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7062b119-01a4-45f4-bb52-212178382271', 'attached_at': '', 'detached_at': '', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'serial': '7ab345f6-458d-4bb0-be22-c593a481ce3f'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 844.891694] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac499728-7ae0-4a69-aa1f-bb19ac824b2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.911030] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ca3df5-8ddb-4e9a-bb99-c9630dfc6e0b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.918830] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96752e53-d642-4c5a-b9a5-9b45a725c4d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.946795] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80b4600-9cba-4496-b130-a7590c7a1501 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.950445] env[65107]: INFO nova.compute.manager [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Took 48.52 seconds to build instance. [ 844.960919] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102856, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.972980] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] The volume has not been displaced from its original location: [datastore2] volume-7ab345f6-458d-4bb0-be22-c593a481ce3f/volume-7ab345f6-458d-4bb0-be22-c593a481ce3f.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 844.978486] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Reconfiguring VM instance instance-00000017 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 844.984192] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2baae1f7-6b50-40b6-a255-2985af1126dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.008158] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098432} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.009783] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.010198] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 845.010198] env[65107]: value = "task-5102859" [ 845.010198] env[65107]: _type = "Task" [ 845.010198] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.010940] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88126244-22f7-451e-bcd0-c15152dd29c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.041922] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 9a46ff51-4035-4709-ae6c-0fb5397200ab/9a46ff51-4035-4709-ae6c-0fb5397200ab.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.050465] env[65107]: INFO nova.compute.manager [-] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Took 1.33 seconds to deallocate network for instance. [ 845.050743] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6a9e3b4-6b5b-4788-b0ad-3eb054ede650 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.066621] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.066833] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102854, 'name': ReconfigVM_Task, 'duration_secs': 0.598268} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.068697] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c/21c471cd-2435-4e4e-82cc-8cdf03f1fa6c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.074372] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e1dd0ac-4bda-4c35-b900-40eeb9cb3086 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.080020] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 845.080020] env[65107]: value = "task-5102860" [ 845.080020] env[65107]: _type = "Task" [ 845.080020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.080020] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 845.080020] env[65107]: value = "task-5102861" [ 845.080020] env[65107]: _type = "Task" [ 845.080020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.100067] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102861, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.100686] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102860, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.278584] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.602s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.282493] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.934s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.282869] env[65107]: DEBUG nova.objects.instance [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lazy-loading 'resources' on Instance uuid 65d249a9-f79c-46c0-8630-169937cbcaf7 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.297989] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.298209] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.311962] env[65107]: INFO nova.scheduler.client.report [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Deleted allocations for instance 1253473b-d4bc-49ac-9724-9aa1e7a8f038 [ 845.351036] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102858, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.352090] env[65107]: DEBUG nova.network.neutron [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 845.376477] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.376783] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.455932] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ea15bf-f8fd-4f0f-9a2f-3c650ce2368d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.035s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.458039] env[65107]: WARNING neutronclient.v2_0.client [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.458804] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.459224] env[65107]: WARNING openstack [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.471543] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102856, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.524505] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102859, 'name': ReconfigVM_Task, 'duration_secs': 0.197199} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.524682] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Reconfigured VM instance instance-00000017 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 845.529498] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-923064d9-db34-49da-8d7f-9c8a6c9cb03f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.554968] env[65107]: DEBUG nova.network.neutron [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updating instance_info_cache with network_info: [{"id": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "address": "fa:16:3e:6b:29:71", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb68bf5-be", "ovs_interfaceid": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 845.559590] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 845.559590] env[65107]: value = "task-5102862" [ 845.559590] env[65107]: _type = "Task" [ 845.559590] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.570915] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.575748] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.594334] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102861, 'name': Rename_Task, 'duration_secs': 0.284358} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.597279] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.597667] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102860, 'name': ReconfigVM_Task, 'duration_secs': 0.29823} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.598235] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a04d739-9e83-4bc7-a6ac-98266732c58c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.600076] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 9a46ff51-4035-4709-ae6c-0fb5397200ab/9a46ff51-4035-4709-ae6c-0fb5397200ab.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.600703] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-054771f5-2684-4828-a390-c5d33c04acc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.608169] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 845.608169] env[65107]: value = "task-5102863" [ 845.608169] env[65107]: _type = "Task" [ 845.608169] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.609873] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 845.609873] env[65107]: value = "task-5102864" [ 845.609873] env[65107]: _type = "Task" [ 845.609873] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.622148] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.626088] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102864, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.823269] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acd4d17a-0933-4823-bbae-2790b99bda50 tempest-AttachInterfacesUnderV243Test-1550338923 tempest-AttachInterfacesUnderV243Test-1550338923-project-member] Lock "1253473b-d4bc-49ac-9724-9aa1e7a8f038" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.920s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.843164] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102858, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.970258] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102856, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.065792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.066168] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Instance network_info: |[{"id": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "address": "fa:16:3e:6b:29:71", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb68bf5-be", "ovs_interfaceid": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 846.066753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:29:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bb68bf5-be11-4086-a291-e8d4eeb2047b', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.074887] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 846.078357] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.081921] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e2196a0-d004-4bb5-975a-d8a1bebfaa44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.096472] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102862, 'name': ReconfigVM_Task, 'duration_secs': 0.271538} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.099371] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992609', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'name': 'volume-7ab345f6-458d-4bb0-be22-c593a481ce3f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7062b119-01a4-45f4-bb52-212178382271', 'attached_at': '', 'detached_at': '', 'volume_id': '7ab345f6-458d-4bb0-be22-c593a481ce3f', 'serial': '7ab345f6-458d-4bb0-be22-c593a481ce3f'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 846.099737] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.101194] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45185b48-9804-44bb-af49-e553a82d9f47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.106351] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.106351] env[65107]: value = "task-5102865" [ 846.106351] env[65107]: _type = "Task" [ 846.106351] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.112755] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.118805] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dff34073-5ced-4b60-a056-59492ead15e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.128357] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102865, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.135540] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102863, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.144260] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102864, 'name': Rename_Task, 'duration_secs': 0.374478} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.144260] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.144260] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2ce8d48-6c4c-492a-9e2d-bcda662acd1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.153625] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 846.153625] env[65107]: value = "task-5102867" [ 846.153625] env[65107]: _type = "Task" [ 846.153625] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.169186] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.214257] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.214535] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.214774] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Deleting the datastore file [datastore2] 7062b119-01a4-45f4-bb52-212178382271 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.215191] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5dbad47-4c27-4e28-9db6-63016596839e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.224450] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for the task: (returnval){ [ 846.224450] env[65107]: value = "task-5102868" [ 846.224450] env[65107]: _type = "Task" [ 846.224450] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.241024] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102868, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.348236] env[65107]: DEBUG oslo_vmware.api [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102858, 'name': RemoveSnapshot_Task, 'duration_secs': 1.346715} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.348519] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 846.348795] env[65107]: INFO nova.compute.manager [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 17.99 seconds to snapshot the instance on the hypervisor. [ 846.451757] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7236596-657f-450b-be19-bd381900b935 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.462357] env[65107]: DEBUG nova.compute.manager [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Received event network-changed-8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 846.462635] env[65107]: DEBUG nova.compute.manager [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Refreshing instance network info cache due to event network-changed-8bb68bf5-be11-4086-a291-e8d4eeb2047b. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 846.462853] env[65107]: DEBUG oslo_concurrency.lockutils [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Acquiring lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.463062] env[65107]: DEBUG oslo_concurrency.lockutils [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Acquired lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.463511] env[65107]: DEBUG nova.network.neutron [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Refreshing network info cache for port 8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 846.469260] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc9839f-7a73-4190-b96b-4052f0fa528f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.478174] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102856, 'name': CreateSnapshot_Task, 'duration_secs': 1.571761} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.479044] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 846.480408] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12b0e3c-2016-4af7-8ce8-5c07a71dfed6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.513612] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe61e0bf-955f-4203-bd87-eae079ae1f70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.529768] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e904bfe8-b175-400f-80d2-47dea8e6138c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.547179] env[65107]: DEBUG nova.compute.provider_tree [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.619147] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102865, 'name': CreateVM_Task, 'duration_secs': 0.383309} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.619855] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.620354] env[65107]: WARNING neutronclient.v2_0.client [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.620937] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.620937] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.621994] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 846.621994] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e21535f5-c177-470d-a4e8-34bfb2abb872 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.626452] env[65107]: DEBUG oslo_vmware.api [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102863, 'name': PowerOnVM_Task, 'duration_secs': 0.826558} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.626952] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.627168] env[65107]: INFO nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Took 9.35 seconds to spawn the instance on the hypervisor. [ 846.627358] env[65107]: DEBUG nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 846.628283] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086e790e-38c4-4e25-ade2-d6f494371c9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.632484] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 846.632484] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5200bd54-1673-838e-4f75-7d3558849146" [ 846.632484] env[65107]: _type = "Task" [ 846.632484] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.650835] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5200bd54-1673-838e-4f75-7d3558849146, 'name': SearchDatastore_Task, 'duration_secs': 0.014422} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.650835] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.650938] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.651154] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.651276] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.651451] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.651767] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e9e3911-60c5-4c84-a953-feb9df5e167b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.664842] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102867, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.665808] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.665998] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.666793] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fa751ca-043b-4d2a-b908-84a4e93545e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.674593] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 846.674593] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527bd39e-c658-3e6f-2e8c-e11b139950a7" [ 846.674593] env[65107]: _type = "Task" [ 846.674593] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.686724] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527bd39e-c658-3e6f-2e8c-e11b139950a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.735678] env[65107]: DEBUG oslo_vmware.api [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Task: {'id': task-5102868, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104173} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.735984] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.736199] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.736406] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.736601] env[65107]: INFO nova.compute.manager [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Took 2.37 seconds to destroy the instance on the hypervisor. [ 846.736865] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 846.737099] env[65107]: DEBUG nova.compute.manager [-] [instance: 7062b119-01a4-45f4-bb52-212178382271] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 846.737211] env[65107]: DEBUG nova.network.neutron [-] [instance: 7062b119-01a4-45f4-bb52-212178382271] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 846.737488] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.738110] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.738401] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.802336] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.904718] env[65107]: DEBUG nova.compute.manager [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Found 3 images (rotation: 2) {{(pid=65107) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 846.904718] env[65107]: DEBUG nova.compute.manager [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Rotating out 1 backups {{(pid=65107) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5082}} [ 846.904718] env[65107]: DEBUG nova.compute.manager [None req-c790274e-f02a-4738-9dc1-3df6e8bd8466 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleting image 7cb99544-c916-4db5-b034-88926221b037 {{(pid=65107) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5087}} [ 846.969121] env[65107]: WARNING neutronclient.v2_0.client [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.972027] env[65107]: WARNING openstack [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.972027] env[65107]: WARNING openstack [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.034608] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 847.035401] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7f203af3-1530-460a-9929-8bf0c5db787b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.049846] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 847.049846] env[65107]: value = "task-5102869" [ 847.049846] env[65107]: _type = "Task" [ 847.049846] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.053990] env[65107]: DEBUG nova.scheduler.client.report [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.065750] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102869, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.158487] env[65107]: INFO nova.compute.manager [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Took 49.93 seconds to build instance. [ 847.165887] env[65107]: DEBUG oslo_vmware.api [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102867, 'name': PowerOnVM_Task, 'duration_secs': 0.559419} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.166221] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.166541] env[65107]: INFO nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Took 7.21 seconds to spawn the instance on the hypervisor. [ 847.166758] env[65107]: DEBUG nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 847.168029] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27363006-673a-4f01-b758-171c1adfe544 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.193528] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527bd39e-c658-3e6f-2e8c-e11b139950a7, 'name': SearchDatastore_Task, 'duration_secs': 0.014156} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.195428] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46610c12-52a4-40aa-b2d6-41e4ebe75864 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.202527] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 847.202527] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5234081f-c405-f5fd-db97-a5ff22b053eb" [ 847.202527] env[65107]: _type = "Task" [ 847.202527] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.212313] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5234081f-c405-f5fd-db97-a5ff22b053eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.280981] env[65107]: WARNING openstack [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.281424] env[65107]: WARNING openstack [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.388275] env[65107]: WARNING neutronclient.v2_0.client [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 847.388275] env[65107]: WARNING openstack [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.388275] env[65107]: WARNING openstack [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.562453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.280s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.564695] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102869, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.565480] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 43.673s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.565590] env[65107]: DEBUG nova.objects.instance [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lazy-loading 'resources' on Instance uuid f2814dd6-8be5-49b2-a030-f57e452c2e6d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.594275] env[65107]: INFO nova.scheduler.client.report [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Deleted allocations for instance 65d249a9-f79c-46c0-8630-169937cbcaf7 [ 847.661723] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3945bb84-d318-45c4-93a2-c2584a75416d tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.445s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.690761] env[65107]: INFO nova.compute.manager [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Took 49.56 seconds to build instance. [ 847.714994] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5234081f-c405-f5fd-db97-a5ff22b053eb, 'name': SearchDatastore_Task, 'duration_secs': 0.017605} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.715473] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 847.715573] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 701a1a16-9348-4748-bb24-ee245a76566c/701a1a16-9348-4748-bb24-ee245a76566c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.715836] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8ed02f6-6c40-4526-9e77-4fdc484d33c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.723954] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 847.723954] env[65107]: value = "task-5102870" [ 847.723954] env[65107]: _type = "Task" [ 847.723954] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.736702] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.967250] env[65107]: DEBUG nova.network.neutron [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updated VIF entry in instance network info cache for port 8bb68bf5-be11-4086-a291-e8d4eeb2047b. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 847.967798] env[65107]: DEBUG nova.network.neutron [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updating instance_info_cache with network_info: [{"id": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "address": "fa:16:3e:6b:29:71", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb68bf5-be", "ovs_interfaceid": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.063186] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102869, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.109316] env[65107]: DEBUG oslo_concurrency.lockutils [None req-825a80e7-6919-4073-bc25-46d9fe945fed tempest-ServerShowV257Test-143620786 tempest-ServerShowV257Test-143620786-project-member] Lock "65d249a9-f79c-46c0-8630-169937cbcaf7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.997s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.196850] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ccea7e96-56e0-478a-892c-e6a6cb0a7f34 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.073s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.240308] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102870, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.302606] env[65107]: DEBUG nova.network.neutron [-] [instance: 7062b119-01a4-45f4-bb52-212178382271] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.365930] env[65107]: DEBUG nova.compute.manager [req-a4d604e1-a3d7-4ba5-9a6c-d7bed9fcadcf req-89d5b258-a183-4fe2-b260-b274ff46e8ee service nova] [instance: 7062b119-01a4-45f4-bb52-212178382271] Received event network-vif-deleted-09df7af1-1609-4648-a7d1-a870687d5bbb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 848.480053] env[65107]: DEBUG oslo_concurrency.lockutils [req-fdad960f-8f78-4b6b-80e3-c71c265eefe8 req-14e43cd8-bbf5-40a0-bb79-fd8bc778cb14 service nova] Releasing lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.570993] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102869, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.687489] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31b9d8b-8dc3-4699-84cc-72583c7a3a1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.699116] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d7b025-a787-4903-83b7-5ece57e0dda0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.737832] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90574e9c-ca61-41fb-bbc0-126574b5c8c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.755950] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7917feb7-9832-4cd9-80fa-9cb241a21cc6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.765304] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.748888} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.765304] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 701a1a16-9348-4748-bb24-ee245a76566c/701a1a16-9348-4748-bb24-ee245a76566c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.765304] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.765304] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa3fa082-237b-49d5-a1b4-6d5d5deed28b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.777950] env[65107]: DEBUG nova.compute.provider_tree [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.786834] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 848.786834] env[65107]: value = "task-5102871" [ 848.786834] env[65107]: _type = "Task" [ 848.786834] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.796455] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.804786] env[65107]: INFO nova.compute.manager [-] [instance: 7062b119-01a4-45f4-bb52-212178382271] Took 2.07 seconds to deallocate network for instance. [ 849.067698] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102869, 'name': CloneVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.282185] env[65107]: DEBUG nova.scheduler.client.report [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.300715] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074844} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.301089] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.302244] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea820380-660d-49a7-8058-e8ec33baf762 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.329186] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 701a1a16-9348-4748-bb24-ee245a76566c/701a1a16-9348-4748-bb24-ee245a76566c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.329992] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d795bd78-9598-4623-8c5e-b2e33730ca9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.353911] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 849.353911] env[65107]: value = "task-5102872" [ 849.353911] env[65107]: _type = "Task" [ 849.353911] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.366969] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102872, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.368458] env[65107]: INFO nova.compute.manager [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Took 0.56 seconds to detach 1 volumes for instance. [ 849.371977] env[65107]: DEBUG nova.compute.manager [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] [instance: 7062b119-01a4-45f4-bb52-212178382271] Deleting volume: 7ab345f6-458d-4bb0-be22-c593a481ce3f {{(pid=65107) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 849.569833] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102869, 'name': CloneVM_Task, 'duration_secs': 2.154494} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.569833] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Created linked-clone VM from snapshot [ 849.570351] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c6e806-da60-48dc-97a2-8a4a2a0b13fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.585818] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Uploading image 41b9b11a-351c-4992-af7b-f3b9c45ca14e {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 849.618545] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 849.618545] env[65107]: value = "vm-992751" [ 849.618545] env[65107]: _type = "VirtualMachine" [ 849.618545] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 849.618770] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f1680c0b-ded0-4c89-87b9-caaf84133611 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.628373] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease: (returnval){ [ 849.628373] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298bd29-144e-673a-cf8c-88a27e4b9826" [ 849.628373] env[65107]: _type = "HttpNfcLease" [ 849.628373] env[65107]: } obtained for exporting VM: (result){ [ 849.628373] env[65107]: value = "vm-992751" [ 849.628373] env[65107]: _type = "VirtualMachine" [ 849.628373] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 849.628732] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the lease: (returnval){ [ 849.628732] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298bd29-144e-673a-cf8c-88a27e4b9826" [ 849.628732] env[65107]: _type = "HttpNfcLease" [ 849.628732] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 849.638848] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.638848] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298bd29-144e-673a-cf8c-88a27e4b9826" [ 849.638848] env[65107]: _type = "HttpNfcLease" [ 849.638848] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 849.790267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.224s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.792970] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 43.151s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.828463] env[65107]: INFO nova.scheduler.client.report [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Deleted allocations for instance f2814dd6-8be5-49b2-a030-f57e452c2e6d [ 849.873318] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102872, 'name': ReconfigVM_Task, 'duration_secs': 0.363415} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.874053] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 701a1a16-9348-4748-bb24-ee245a76566c/701a1a16-9348-4748-bb24-ee245a76566c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.874697] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80d02aab-53e2-417c-b04f-bcf3eeda3f9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.885397] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 849.885397] env[65107]: value = "task-5102875" [ 849.885397] env[65107]: _type = "Task" [ 849.885397] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.896887] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102875, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.927849] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.137989] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.137989] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298bd29-144e-673a-cf8c-88a27e4b9826" [ 850.137989] env[65107]: _type = "HttpNfcLease" [ 850.137989] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 850.138425] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 850.138425] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5298bd29-144e-673a-cf8c-88a27e4b9826" [ 850.138425] env[65107]: _type = "HttpNfcLease" [ 850.138425] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 850.139278] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3c7885-2700-48a5-8e0a-adf1b828b6e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.148797] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e875da-af84-ba13-2afc-6aae3fdc6afe/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 850.148934] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e875da-af84-ba13-2afc-6aae3fdc6afe/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 850.274161] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1b2e1128-64ce-47b5-9f17-84228e8a0501 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.336705] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9a182bc-0efc-42ad-b4ff-f11a1656d319 tempest-ListImageFiltersTestJSON-812883063 tempest-ListImageFiltersTestJSON-812883063-project-member] Lock "f2814dd6-8be5-49b2-a030-f57e452c2e6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.940s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.396021] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102875, 'name': Rename_Task, 'duration_secs': 0.174407} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.397222] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.397222] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4169a05d-f48a-4c98-824a-af0c4473afa6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.405996] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 850.405996] env[65107]: value = "task-5102876" [ 850.405996] env[65107]: _type = "Task" [ 850.405996] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.422357] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.809340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.809699] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.810257] env[65107]: DEBUG nova.compute.manager [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 850.811322] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Applying migration context for instance 369caf24-48a6-4640-8601-e30fd47453da as it has an incoming, in-progress migration 9e867391-8bde-4c95-983e-919c66e15cd7. Migration status is reverting {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 850.813558] env[65107]: INFO nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating resource usage from migration 9e867391-8bde-4c95-983e-919c66e15cd7 [ 850.816826] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a410dd06-3df9-432e-8c67-55492059cb8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.689902] env[65107]: DEBUG nova.compute.manager [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 851.690947] env[65107]: DEBUG nova.objects.instance [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'flavor' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.698467] env[65107]: DEBUG oslo_vmware.api [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102876, 'name': PowerOnVM_Task, 'duration_secs': 1.264378} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.698586] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.698883] env[65107]: INFO nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Took 7.98 seconds to spawn the instance on the hypervisor. [ 851.699174] env[65107]: DEBUG nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 851.700358] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c45456-34d5-4627-b490-9c7dd7fa1cf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e3510c31-d6be-4e9f-a0a2-a662123861e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b779b5ad-7d93-47c2-b824-6d76246c00f5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.719024] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 8306ff90-b8bd-4270-8133-96abe483156b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 7062b119-01a4-45f4-bb52-212178382271 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 71e6e280-50c3-4655-a6a3-40ea4c650d96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ef417f32-8c0e-4a62-9078-cb2bf95b85f8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d2d0e18c-4839-416c-b363-b93996e8ba7f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bfe544ec-9b08-4118-a940-a51520ecaac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 82e8c093-b718-4d38-9682-ba2e710e4b93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance a4f43fbc-63e3-4b69-bdb1-7aaf47aae090 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Migration 9e867391-8bde-4c95-983e-919c66e15cd7 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 369caf24-48a6-4640-8601-e30fd47453da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719024] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719701] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 179e613b-e5a2-4fbc-8fa6-b72769425ff3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.719701] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 780f910f-4f73-41ba-a795-0daae3097314 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.720029] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 5fccc802-f5ac-4450-8246-4cf9a5371046 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.720029] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 805e2210-fb58-48da-be63-5d548b2f8a2f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 851.720195] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 2d0a93a4-59a8-4788-93ce-c34af1432a4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.720913] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 40474981-eeef-492e-8a8b-aaea7f554c02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.720913] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.720913] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 9a46ff51-4035-4709-ae6c-0fb5397200ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 851.721089] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 701a1a16-9348-4748-bb24-ee245a76566c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 852.230371] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 88cac920-3c1e-41dc-9e52-f51f48a80f56 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 852.236987] env[65107]: INFO nova.compute.manager [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Took 53.71 seconds to build instance. [ 852.710213] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.710878] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75d592ec-d686-489f-b02d-84a543e4f9ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.720098] env[65107]: DEBUG oslo_vmware.api [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 852.720098] env[65107]: value = "task-5102877" [ 852.720098] env[65107]: _type = "Task" [ 852.720098] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.730903] env[65107]: DEBUG oslo_vmware.api [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.733851] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ee50d08a-57fc-4c05-96fe-a11fe2708165 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 852.741433] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2544a4f8-5b4c-41d1-8f66-17f22d8aa8b1 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.223s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.805158] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "40474981-eeef-492e-8a8b-aaea7f554c02" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 852.805537] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.805796] env[65107]: DEBUG nova.compute.manager [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 852.807553] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cea7fef-a62f-4ccb-b333-bf54e3265d5b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.816538] env[65107]: DEBUG nova.compute.manager [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 852.817311] env[65107]: DEBUG nova.objects.instance [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lazy-loading 'flavor' on Instance uuid 40474981-eeef-492e-8a8b-aaea7f554c02 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.232077] env[65107]: DEBUG oslo_vmware.api [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102877, 'name': PowerOffVM_Task, 'duration_secs': 0.259598} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.232248] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.233364] env[65107]: DEBUG nova.compute.manager [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 853.233364] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0294aa1-02db-444f-b153-f5bfdfdbbf4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.241928] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 853.331833] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "701a1a16-9348-4748-bb24-ee245a76566c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.332103] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.332471] env[65107]: INFO nova.compute.manager [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Shelving [ 853.745792] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 40e17da0-b459-4619-8e9e-18c3093328c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 853.751424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6fc0f968-5849-454a-a41b-49e96617647d tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.941s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.829428] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.829636] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef0d315e-49dc-4cdd-b4fb-3ea780c3134d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.839161] env[65107]: DEBUG oslo_vmware.api [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 853.839161] env[65107]: value = "task-5102878" [ 853.839161] env[65107]: _type = "Task" [ 853.839161] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.854591] env[65107]: DEBUG oslo_vmware.api [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.251806] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3f34d6a2-0370-4239-9556-b4333e454d58 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 854.350279] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 854.350447] env[65107]: DEBUG oslo_vmware.api [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102878, 'name': PowerOffVM_Task, 'duration_secs': 0.298211} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.350999] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4117119-3c28-4762-88e2-320d2edd044f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.352361] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.352545] env[65107]: DEBUG nova.compute.manager [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 854.353339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3309b4a-7e7f-4831-a0ee-4dc21d4e894d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.363985] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 854.363985] env[65107]: value = "task-5102879" [ 854.363985] env[65107]: _type = "Task" [ 854.363985] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.376302] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.755091] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 9fa949ea-358c-46e7-a0f2-4c3275493b64 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 854.755558] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 854.755646] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3776MB phys_disk=100GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '15', 'num_vm_active': '12', 'num_task_None': '12', 'num_os_type_None': '15', 'num_proj_a2d4844feda0432ebe1ef160b0070b83': '1', 'io_workload': '1', 'num_proj_2882c2c7949045519a146655bb694a3f': '2', 'num_proj_f379144b78764fe394039d87b043a946': '1', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_71cd673a320348a9925e8f59637afc17': '1', 'num_proj_e1fd7913632a4a5b9e076b5b156fa5fa': '1', 'num_proj_544c649453844164bac53c98b342543f': '1', 'num_proj_634f89d91e1443c7a0eb1314ecb17232': '1', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '1', 'num_proj_b981c5f6b3b64476a2406cd9549f7d6b': '1', 'num_vm_suspended': '1', 'num_task_image_uploading': '1', 'num_proj_6f9188f3e35e4d06a977e2180918c616': '1', 'num_proj_f883a2e25da34a548b020184acb9245e': '3', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_4778fe9152224fd29d1f6220a19b5a36': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 854.875903] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1cfe4e99-b0bf-47bb-a84a-da8ce39d5df0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.070s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.881513] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102879, 'name': PowerOffVM_Task, 'duration_secs': 0.220994} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.882211] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.882682] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4356e0b-0f74-4261-85a1-1ff7c02c98dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.909484] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20a5ce1-03fe-4105-96a5-ef9e645f71ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.153220] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.153362] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.288911] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696a02b0-c5d9-4c72-b586-0931500d69d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.297858] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504e1bbf-2ced-4a2e-bade-15f34a1c1935 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.329855] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0653ecea-a2f7-45b0-9ed1-9edc8a28af4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.338688] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601a0d61-1cda-4fc5-9e0a-7d1b2c1ac0cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.353315] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.425245] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 855.425573] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d45f9bf9-e044-4c3f-8b3b-dc7532487f29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.435471] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 855.435471] env[65107]: value = "task-5102880" [ 855.435471] env[65107]: _type = "Task" [ 855.435471] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.445618] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102880, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.656159] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 855.857111] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.948062] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102880, 'name': CreateSnapshot_Task, 'duration_secs': 0.458766} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.948404] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 855.949188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e289502-376f-4a1d-95d4-213c4e5a7f98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.125485] env[65107]: DEBUG nova.objects.instance [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lazy-loading 'flavor' on Instance uuid 40474981-eeef-492e-8a8b-aaea7f554c02 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.194110] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.363274] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 856.363274] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.571s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.363485] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 46.572s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.363534] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.365818] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 41.601s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.368053] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.368053] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Cleaning up deleted instances {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 856.391394] env[65107]: INFO nova.scheduler.client.report [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Deleted allocations for instance 8306ff90-b8bd-4270-8133-96abe483156b [ 856.471085] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 856.472119] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8116b839-c615-450e-9732-8490a9ebaee6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.481797] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 856.481797] env[65107]: value = "task-5102881" [ 856.481797] env[65107]: _type = "Task" [ 856.481797] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.491363] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102881, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.559499] env[65107]: DEBUG nova.compute.manager [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Stashing vm_state: stopped {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 856.632995] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.633496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquired lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 856.633496] env[65107]: DEBUG nova.network.neutron [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 856.633590] env[65107]: DEBUG nova.objects.instance [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lazy-loading 'info_cache' on Instance uuid 40474981-eeef-492e-8a8b-aaea7f554c02 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.871134] env[65107]: DEBUG nova.objects.instance [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lazy-loading 'migration_context' on Instance uuid 369caf24-48a6-4640-8601-e30fd47453da {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.882046] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] There are 35 instances to clean {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 856.882237] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 65d249a9-f79c-46c0-8630-169937cbcaf7] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 856.901968] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d8de6a90-562f-4ca7-b89a-e4d891d955e1 tempest-VolumesAssistedSnapshotsTest-997297769 tempest-VolumesAssistedSnapshotsTest-997297769-project-member] Lock "8306ff90-b8bd-4270-8133-96abe483156b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 51.098s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.994649] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102881, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.081219] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.139242] env[65107]: DEBUG nova.objects.base [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Object Instance<40474981-eeef-492e-8a8b-aaea7f554c02> lazy-loaded attributes: flavor,info_cache {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 857.388859] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 5334a3dc-e0b5-43e4-b1fb-4c643aac4f5c] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 857.500640] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102881, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.642568] env[65107]: WARNING neutronclient.v2_0.client [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.643329] env[65107]: WARNING openstack [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 857.643736] env[65107]: WARNING openstack [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 857.878295] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66da175-cc91-46ca-9f9a-d14a330eefbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.887355] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac63756-8ec3-44bb-9338-ba86f527fbfd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.891835] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 19518d90-ba6f-44e5-b7c7-1ececad2b54d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 857.927608] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37472c6f-e1d7-45c0-ac94-5809733d937f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.937770] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e07102-ff07-4bd6-abd2-384bb2504d38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.954386] env[65107]: DEBUG nova.compute.provider_tree [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.996160] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102881, 'name': CloneVM_Task, 'duration_secs': 1.159375} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.996790] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Created linked-clone VM from snapshot [ 857.997728] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ae476c-667d-4062-b799-2d6994cc5454 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.007401] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Uploading image 4946aa6f-8b8a-4d0a-955a-55a2bdb46801 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 858.046096] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 858.046096] env[65107]: value = "vm-992753" [ 858.046096] env[65107]: _type = "VirtualMachine" [ 858.046096] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 858.046096] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fe15d009-d5a1-4edc-bc64-69fb53452702 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.053721] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lease: (returnval){ [ 858.053721] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a274e-204c-4cd6-dee6-a6e3b7aa84c4" [ 858.053721] env[65107]: _type = "HttpNfcLease" [ 858.053721] env[65107]: } obtained for exporting VM: (result){ [ 858.053721] env[65107]: value = "vm-992753" [ 858.053721] env[65107]: _type = "VirtualMachine" [ 858.053721] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 858.054471] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the lease: (returnval){ [ 858.054471] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a274e-204c-4cd6-dee6-a6e3b7aa84c4" [ 858.054471] env[65107]: _type = "HttpNfcLease" [ 858.054471] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 858.063662] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.063662] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a274e-204c-4cd6-dee6-a6e3b7aa84c4" [ 858.063662] env[65107]: _type = "HttpNfcLease" [ 858.063662] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 858.397775] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 6563cbda-5fd3-4640-9e9a-95904f3f8335] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 858.462484] env[65107]: DEBUG nova.scheduler.client.report [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 858.565319] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.565319] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a274e-204c-4cd6-dee6-a6e3b7aa84c4" [ 858.565319] env[65107]: _type = "HttpNfcLease" [ 858.565319] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 858.565319] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 858.565319] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a274e-204c-4cd6-dee6-a6e3b7aa84c4" [ 858.565319] env[65107]: _type = "HttpNfcLease" [ 858.565319] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 858.565969] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8cd5d7-89ae-4902-b67d-5d824d3d42f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.577412] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c8fefc-57c1-6114-7073-ba6d5260b95c/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 858.581966] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c8fefc-57c1-6114-7073-ba6d5260b95c/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 858.690844] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1033d07e-8083-437c-bce7-9409e4f9a1b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.900255] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: c52b1f08-4000-4ad1-bc41-c1d0a41dcd2d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 859.023123] env[65107]: WARNING openstack [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.023682] env[65107]: WARNING openstack [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.247766] env[65107]: WARNING neutronclient.v2_0.client [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 859.248942] env[65107]: WARNING openstack [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.249595] env[65107]: WARNING openstack [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.297856] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e875da-af84-ba13-2afc-6aae3fdc6afe/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 859.299166] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5c8de2-2496-4ab9-9f5c-d92963b99744 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.307044] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e875da-af84-ba13-2afc-6aae3fdc6afe/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 859.307829] env[65107]: ERROR oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e875da-af84-ba13-2afc-6aae3fdc6afe/disk-0.vmdk due to incomplete transfer. [ 859.307829] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-58873252-7e0e-4b09-bef3-85abd1666c04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.317790] env[65107]: DEBUG oslo_vmware.rw_handles [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e875da-af84-ba13-2afc-6aae3fdc6afe/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 859.318821] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Uploaded image 41b9b11a-351c-4992-af7b-f3b9c45ca14e to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 859.321496] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 859.322328] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-67277d79-9ac1-4ef8-8af4-2505eaf93692 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.334674] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 859.334674] env[65107]: value = "task-5102883" [ 859.334674] env[65107]: _type = "Task" [ 859.334674] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.351333] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102883, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.409635] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d40aec5a-d898-42f3-a121-503cf043f72b] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 859.443123] env[65107]: DEBUG nova.network.neutron [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Updating instance_info_cache with network_info: [{"id": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "address": "fa:16:3e:8e:07:f3", "network": {"id": "1cc5ef01-651d-4b44-8e7d-5cea35f4d9cb", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-568216530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f883a2e25da34a548b020184acb9245e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860d5e06-e0", "ovs_interfaceid": "860d5e06-e063-471a-8b43-cda8c6bd94b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 859.475342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.109s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.484640] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 43.907s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.486700] env[65107]: INFO nova.compute.claims [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.848660] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102883, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.913117] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 854f7a3f-9b7a-4506-8310-94533f0e4cfc] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 859.948849] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Releasing lock "refresh_cache-40474981-eeef-492e-8a8b-aaea7f554c02" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 860.348446] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102883, 'name': Destroy_Task, 'duration_secs': 0.700842} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.348756] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Destroyed the VM [ 860.348986] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 860.349272] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0620abc4-fc43-49d1-bd8b-b50fe8cf3173 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.356690] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 860.356690] env[65107]: value = "task-5102884" [ 860.356690] env[65107]: _type = "Task" [ 860.356690] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.365352] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102884, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.419307] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 90706f49-af87-4038-ae76-02a95e4a12e6] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 860.873312] env[65107]: DEBUG oslo_vmware.api [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102884, 'name': RemoveSnapshot_Task, 'duration_secs': 0.450428} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.873858] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 860.873858] env[65107]: INFO nova.compute.manager [None req-438fc3f1-d1fb-4f32-b106-06301ae7a3a6 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Took 17.01 seconds to snapshot the instance on the hypervisor. [ 860.923926] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: c77e0e6c-7e5c-45a7-a0f3-d10ee3f920e0] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 860.967455] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.967455] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-989b9d02-ba44-4797-97d3-c47818fa50b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.975562] env[65107]: DEBUG oslo_vmware.api [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 860.975562] env[65107]: value = "task-5102885" [ 860.975562] env[65107]: _type = "Task" [ 860.975562] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.991188] env[65107]: DEBUG oslo_vmware.api [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102885, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.043961] env[65107]: INFO nova.compute.manager [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Swapping old allocation on dict_keys(['8adccf8d-6d73-43c0-ad15-ee89967ab8cd']) held by migration 9e867391-8bde-4c95-983e-919c66e15cd7 for instance [ 861.080320] env[65107]: DEBUG nova.scheduler.client.report [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Overwriting current allocation {'allocations': {'8adccf8d-6d73-43c0-ad15-ee89967ab8cd': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 88}}, 'project_id': '2882c2c7949045519a146655bb694a3f', 'user_id': '66f11f09dd98447bb9851bee92183dcc', 'consumer_generation': 1} on consumer 369caf24-48a6-4640-8601-e30fd47453da {{(pid=65107) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 861.150042] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.246672] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e726184-7a4a-4dc6-8636-7bae0afc2d60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.259524] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360213e8-2c35-4fc2-91b2-2e78dddff711 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.302709] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff645d3-8536-4606-a1a5-2666cbbefc28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.312741] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0320ad47-6875-49dc-8fa7-ab39d7411fd5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.321018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.321018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquired lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.321018] env[65107]: DEBUG nova.network.neutron [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 861.332507] env[65107]: DEBUG nova.compute.provider_tree [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.423082] env[65107]: INFO nova.compute.manager [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Rebuilding instance [ 861.427958] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3948dcdd-e7c1-40b6-a3ec-97741940a0fb] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 861.488389] env[65107]: DEBUG oslo_vmware.api [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102885, 'name': PowerOnVM_Task, 'duration_secs': 0.480542} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.493354] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.493354] env[65107]: DEBUG nova.compute.manager [None req-5814871a-6026-4a58-94a0-3439a476e92c tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 861.493789] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0a9e5e-d04c-4709-ba43-7c0c62d524f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.501540] env[65107]: DEBUG nova.compute.manager [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 861.503089] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc7028c-70d3-4223-ba98-59d1c6defa0c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.824443] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.826189] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.826189] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.836055] env[65107]: DEBUG nova.scheduler.client.report [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 861.977223] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: c7739dfd-88d6-46ac-b3a0-e7dfe9b00332] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 861.977223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.977223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.977223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.977223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.977223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.977223] env[65107]: INFO nova.compute.manager [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Terminating instance [ 862.051729] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.052223] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.298760] env[65107]: WARNING neutronclient.v2_0.client [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.298760] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.299113] env[65107]: WARNING openstack [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.342089] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.858s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.343083] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 862.346142] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.508s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.347461] env[65107]: INFO nova.compute.claims [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.440698] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 23aac5d2-f8ba-4fc4-a3ce-04e2e307a9b6] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 862.455400] env[65107]: DEBUG nova.compute.manager [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 862.455717] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 862.456585] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc9ef07-40c6-48b8-bba9-0d92f8871365 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.465203] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 862.465500] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64803895-a96f-4f94-9be1-9d16f06635f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.523220] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.523580] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18dae137-53af-4275-bb06-ce093b9b9d74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.530911] env[65107]: DEBUG nova.network.neutron [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [{"id": "fcca9afe-edba-479c-93f6-1310627d946f", "address": "fa:16:3e:49:42:44", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.164", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcca9afe-ed", "ovs_interfaceid": "fcca9afe-edba-479c-93f6-1310627d946f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 862.543032] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 862.543032] env[65107]: value = "task-5102887" [ 862.543032] env[65107]: _type = "Task" [ 862.543032] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.555478] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.556621] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 862.556867] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 862.557062] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleting the datastore file [datastore1] 2d0a93a4-59a8-4788-93ce-c34af1432a4f {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.558308] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fa1281d-610a-4319-8a1b-a57c44f252fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.566203] env[65107]: DEBUG oslo_vmware.api [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 862.566203] env[65107]: value = "task-5102888" [ 862.566203] env[65107]: _type = "Task" [ 862.566203] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.580245] env[65107]: DEBUG oslo_vmware.api [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.851963] env[65107]: DEBUG nova.compute.utils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 862.853690] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 862.853942] env[65107]: DEBUG nova.network.neutron [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 862.854253] env[65107]: WARNING neutronclient.v2_0.client [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.854898] env[65107]: WARNING neutronclient.v2_0.client [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.855536] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.855536] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.945541] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 07b83b81-9c65-4180-b208-c96525b78d8a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 862.969842] env[65107]: DEBUG nova.policy [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '519c7d7fb4f548538c52dcba20b4a436', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25b948bb3ec94c9886bd2fee74004db1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 863.036929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Releasing lock "refresh_cache-369caf24-48a6-4640-8601-e30fd47453da" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.039706] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.040178] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28540aa1-012b-4969-bef8-cff85e09555b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.056488] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102887, 'name': PowerOffVM_Task, 'duration_secs': 0.290705} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.061030] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.061030] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.061030] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 863.061030] env[65107]: value = "task-5102889" [ 863.061030] env[65107]: _type = "Task" [ 863.061030] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.061409] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d2d7fc8-e3b6-4383-a1a0-e2f0ba557748 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.074410] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 863.074410] env[65107]: value = "task-5102890" [ 863.074410] env[65107]: _type = "Task" [ 863.074410] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.082767] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.086494] env[65107]: DEBUG oslo_vmware.api [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335226} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.087293] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.087509] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 863.087736] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.088329] env[65107]: INFO nova.compute.manager [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Took 0.63 seconds to destroy the instance on the hypervisor. [ 863.088542] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 863.088813] env[65107]: DEBUG nova.compute.manager [-] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 863.088972] env[65107]: DEBUG nova.network.neutron [-] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 863.089333] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.090800] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.090800] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.105443] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 863.105670] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 863.105893] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992651', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'name': 'volume-f04439cd-6130-4943-97ee-01d8501c2efc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd65e98bb-f54d-4803-95ca-0fc9b6c184bc', 'attached_at': '', 'detached_at': '', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'serial': 'f04439cd-6130-4943-97ee-01d8501c2efc'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 863.106774] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbceb9a0-5706-47a7-a0a9-f185d9ce7543 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.133534] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fab7cad-dc4c-4d00-ae6e-65176c977233 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.145309] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d39d63-c247-4b09-a433-de1f367b5009 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.172845] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29190a85-67e3-40de-ab6d-66a7e48c4ac3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.195377] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] The volume has not been displaced from its original location: [datastore1] volume-f04439cd-6130-4943-97ee-01d8501c2efc/volume-f04439cd-6130-4943-97ee-01d8501c2efc.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 863.201738] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Reconfiguring VM instance instance-00000028 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 863.202850] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17b0b858-90e7-485d-80ab-28f3b495b2c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.224549] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 863.224549] env[65107]: value = "task-5102891" [ 863.224549] env[65107]: _type = "Task" [ 863.224549] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.235273] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102891, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.236656] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.364715] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 863.385263] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.385263] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.441290] env[65107]: DEBUG nova.network.neutron [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Successfully created port: 2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 863.452992] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: b20a2c71-514d-488a-8edb-8fb0274a894a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 863.579362] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102889, 'name': PowerOffVM_Task, 'duration_secs': 0.258034} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.579362] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.579362] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:22:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a3c77cac-4e5e-476c-bee7-e5d40e2e12d5',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-991042522',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 863.579706] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 863.580485] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 863.580601] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 863.580781] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 863.580928] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 863.581157] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.581320] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 863.581477] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 863.581661] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 863.581834] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 863.594573] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b47410c-fb14-4554-a093-8a04b608598e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.614853] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 863.614853] env[65107]: value = "task-5102892" [ 863.614853] env[65107]: _type = "Task" [ 863.614853] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.631303] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102892, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.680266] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.681500] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.681500] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.681500] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.681725] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.684102] env[65107]: INFO nova.compute.manager [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Terminating instance [ 863.725539] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "ac631507-8abf-4019-bdd5-7684dd51d1e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.726034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.745199] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102891, 'name': ReconfigVM_Task, 'duration_secs': 0.250033} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.745637] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Reconfigured VM instance instance-00000028 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 863.760913] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6687816-5dee-4ca8-a0e0-ba38b78f9305 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.777533] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 863.777533] env[65107]: value = "task-5102893" [ 863.777533] env[65107]: _type = "Task" [ 863.777533] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.793556] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102893, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.887873] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 863.958034] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: a2b72bac-6806-4b7b-b972-32aea52f7c82] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 864.012651] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f371f2b-0622-4faf-a0a2-4a3126ff918c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.022588] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddb8d33-8286-4d4e-86d6-160579b5fda7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.062940] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72440fcc-3d7d-4fcd-bf55-256d900e5594 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.071900] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6371218-4594-4ade-81d1-4ed5f1cb8ab1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.086811] env[65107]: DEBUG nova.compute.provider_tree [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.125411] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102892, 'name': ReconfigVM_Task, 'duration_secs': 0.198113} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.126339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba77c5e-3716-4437-a611-34b784da1d00 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.146214] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:22:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='a3c77cac-4e5e-476c-bee7-e5d40e2e12d5',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-991042522',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 864.146503] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 864.146663] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 864.146846] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 864.146989] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 864.147147] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 864.147347] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.147501] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 864.147667] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 864.147826] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 864.147994] env[65107]: DEBUG nova.virt.hardware [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 864.148853] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dbf4b90-301b-4b15-88a1-d2f6c7925a49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.155073] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 864.155073] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b830b6-7d64-530d-0e56-90ab3eafe87c" [ 864.155073] env[65107]: _type = "Task" [ 864.155073] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.164617] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b830b6-7d64-530d-0e56-90ab3eafe87c, 'name': SearchDatastore_Task} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.170173] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfiguring VM instance instance-00000027 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 864.170390] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38560a20-f767-431d-a7b5-f5f8918bb564 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.190189] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 864.190189] env[65107]: value = "task-5102894" [ 864.190189] env[65107]: _type = "Task" [ 864.190189] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.193729] env[65107]: DEBUG nova.compute.manager [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 864.193930] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.197468] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fee93ea-10c6-4137-bf9b-afdafe05437f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.206109] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102894, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.208985] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.209274] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03183892-3076-4791-92e4-af82707ef369 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.215929] env[65107]: DEBUG oslo_vmware.api [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 864.215929] env[65107]: value = "task-5102895" [ 864.215929] env[65107]: _type = "Task" [ 864.215929] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.226049] env[65107]: DEBUG oslo_vmware.api [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102895, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.236275] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 864.284863] env[65107]: DEBUG nova.network.neutron [-] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.290131] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102893, 'name': ReconfigVM_Task, 'duration_secs': 0.131046} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.290750] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992651', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'name': 'volume-f04439cd-6130-4943-97ee-01d8501c2efc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd65e98bb-f54d-4803-95ca-0fc9b6c184bc', 'attached_at': '', 'detached_at': '', 'volume_id': 'f04439cd-6130-4943-97ee-01d8501c2efc', 'serial': 'f04439cd-6130-4943-97ee-01d8501c2efc'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 864.291193] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.292125] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa320aff-4a2e-435b-b903-5ded199c3401 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.300611] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.300976] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36d39313-b34a-428b-8f52-ab4e6df0d495 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.376120] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 864.392890] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.393253] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.394225] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Deleting the datastore file [datastore1] d65e98bb-f54d-4803-95ca-0fc9b6c184bc {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.395157] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34c69a02-6728-4e78-9862-725d3a41a58f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.406784] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for the task: (returnval){ [ 864.406784] env[65107]: value = "task-5102897" [ 864.406784] env[65107]: _type = "Task" [ 864.406784] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.422199] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102897, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.425218] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 864.425218] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 864.425218] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 864.425532] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 864.425532] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 864.426493] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 864.426493] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.426493] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 864.426493] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 864.426493] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 864.426801] env[65107]: DEBUG nova.virt.hardware [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 864.428104] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ac042c-e3fe-406c-b3c2-b5529d536f80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.433121] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.440046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2719cd-96dc-44bb-ae9f-4995deb356db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.461796] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: ad53146e-210d-4321-b819-90795c1af260] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 864.590849] env[65107]: DEBUG nova.scheduler.client.report [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 864.701887] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.726572] env[65107]: DEBUG oslo_vmware.api [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102895, 'name': PowerOffVM_Task, 'duration_secs': 0.304854} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.726572] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.726755] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.727318] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f301fe42-8caa-43b3-af6d-961b2ce6dbe4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.757715] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.792024] env[65107]: INFO nova.compute.manager [-] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Took 1.70 seconds to deallocate network for instance. [ 864.798234] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.798538] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.798754] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Deleting the datastore file [datastore2] a4f43fbc-63e3-4b69-bdb1-7aaf47aae090 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.799039] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-769c6c5c-546f-419b-8928-59d97dbea2ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.812387] env[65107]: DEBUG oslo_vmware.api [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for the task: (returnval){ [ 864.812387] env[65107]: value = "task-5102899" [ 864.812387] env[65107]: _type = "Task" [ 864.812387] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.820333] env[65107]: DEBUG oslo_vmware.api [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.835204] env[65107]: DEBUG nova.compute.manager [req-9fc03035-2208-4c07-8dcc-bad7fb8fb7d6 req-b914f1f9-bc69-4d23-ab5c-89d2d084ee16 service nova] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Received event network-vif-deleted-acc9d6b8-be91-4755-a23c-2c530ae497fe {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 864.921280] env[65107]: DEBUG oslo_vmware.api [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Task: {'id': task-5102897, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140948} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.921555] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.921714] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.922033] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.969233] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f7446d76-45c2-4e8b-981d-d37c230cf125] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 865.042353] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 865.042877] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef3feb51-50b8-4678-a605-eaa8d7f9d3ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.056699] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e77158-93f0-40b9-aff1-c719ba326b40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.100272] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.753s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.100272] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 865.105381] env[65107]: ERROR nova.compute.manager [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Failed to detach volume f04439cd-6130-4943-97ee-01d8501c2efc from /dev/sda: nova.exception.InstanceNotFound: Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc could not be found. [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Traceback (most recent call last): [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self.driver.rebuild(**kwargs) [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] raise NotImplementedError() [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] NotImplementedError [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] During handling of the above exception, another exception occurred: [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Traceback (most recent call last): [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self.driver.detach_volume(context, old_connection_info, [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] return self._volumeops.detach_volume(connection_info, instance) [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._detach_volume_vmdk(connection_info, instance) [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] stable_ref.fetch_moref(session) [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] nova.exception.InstanceNotFound: Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc could not be found. [ 865.105381] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.116661] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.781s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.116661] env[65107]: INFO nova.compute.claims [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.206614] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102894, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.306287] env[65107]: DEBUG nova.compute.utils [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Build of instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc aborted: Failed to rebuild volume backed instance. {{(pid=65107) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 865.309573] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.311037] env[65107]: ERROR nova.compute.manager [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc aborted: Failed to rebuild volume backed instance. [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Traceback (most recent call last): [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self.driver.rebuild(**kwargs) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] raise NotImplementedError() [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] NotImplementedError [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] During handling of the above exception, another exception occurred: [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Traceback (most recent call last): [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3699, in _rebuild_volume_backed_instance [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._detach_root_volume(context, instance, root_bdm) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3678, in _detach_root_volume [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] with excutils.save_and_reraise_exception(): [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self.force_reraise() [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] raise self.value [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self.driver.detach_volume(context, old_connection_info, [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] return self._volumeops.detach_volume(connection_info, instance) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._detach_volume_vmdk(connection_info, instance) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] stable_ref.fetch_moref(session) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] raise exception.InstanceNotFound(instance_id=self._uuid) [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] nova.exception.InstanceNotFound: Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc could not be found. [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] During handling of the above exception, another exception occurred: [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Traceback (most recent call last): [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 11570, in _error_out_instance_on_exception [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] yield [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3967, in rebuild_instance [ 865.311037] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._do_rebuild_instance_with_claim( [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 4053, in _do_rebuild_instance_with_claim [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._do_rebuild_instance( [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 4245, in _do_rebuild_instance [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._rebuild_default_impl(**kwargs) [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3822, in _rebuild_default_impl [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] self._rebuild_volume_backed_instance( [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] File "/opt/stack/nova/nova/compute/manager.py", line 3714, in _rebuild_volume_backed_instance [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] raise exception.BuildAbortException( [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] nova.exception.BuildAbortException: Build of instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc aborted: Failed to rebuild volume backed instance. [ 865.312639] env[65107]: ERROR nova.compute.manager [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] [ 865.325018] env[65107]: DEBUG oslo_vmware.api [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Task: {'id': task-5102899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195194} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.325325] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.325523] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.325697] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.325871] env[65107]: INFO nova.compute.manager [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Took 1.13 seconds to destroy the instance on the hypervisor. [ 865.326993] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 865.326993] env[65107]: DEBUG nova.compute.manager [-] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 865.326993] env[65107]: DEBUG nova.network.neutron [-] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 865.326993] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.327582] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.327582] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.351330] env[65107]: DEBUG nova.network.neutron [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Successfully updated port: 2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 865.423878] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.472387] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 0e6c9f2d-4412-4d8d-8348-6552a34ab349] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 865.575855] env[65107]: DEBUG nova.compute.manager [req-65ddcbcc-a269-4b56-bf1a-e71cd56bed6e req-336182b7-8eeb-4d52-ad08-7653e5e115f9 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Received event network-vif-plugged-2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 865.576128] env[65107]: DEBUG oslo_concurrency.lockutils [req-65ddcbcc-a269-4b56-bf1a-e71cd56bed6e req-336182b7-8eeb-4d52-ad08-7653e5e115f9 service nova] Acquiring lock "88cac920-3c1e-41dc-9e52-f51f48a80f56-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.576347] env[65107]: DEBUG oslo_concurrency.lockutils [req-65ddcbcc-a269-4b56-bf1a-e71cd56bed6e req-336182b7-8eeb-4d52-ad08-7653e5e115f9 service nova] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.576520] env[65107]: DEBUG oslo_concurrency.lockutils [req-65ddcbcc-a269-4b56-bf1a-e71cd56bed6e req-336182b7-8eeb-4d52-ad08-7653e5e115f9 service nova] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.576819] env[65107]: DEBUG nova.compute.manager [req-65ddcbcc-a269-4b56-bf1a-e71cd56bed6e req-336182b7-8eeb-4d52-ad08-7653e5e115f9 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] No waiting events found dispatching network-vif-plugged-2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 865.577410] env[65107]: WARNING nova.compute.manager [req-65ddcbcc-a269-4b56-bf1a-e71cd56bed6e req-336182b7-8eeb-4d52-ad08-7653e5e115f9 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Received unexpected event network-vif-plugged-2f395da6-1303-4770-b9a5-eed2e9141181 for instance with vm_state building and task_state spawning. [ 865.612188] env[65107]: DEBUG nova.compute.utils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 865.613756] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 865.613960] env[65107]: DEBUG nova.network.neutron [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 865.614304] env[65107]: WARNING neutronclient.v2_0.client [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.614684] env[65107]: WARNING neutronclient.v2_0.client [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.615283] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.615639] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.700448] env[65107]: DEBUG nova.policy [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61d0444710414b8eae5e40ed898d8712', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb7fca6907c4c32a1aa6b8023c39afb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 865.709277] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102894, 'name': ReconfigVM_Task, 'duration_secs': 1.179289} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.709577] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfigured VM instance instance-00000027 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 865.710405] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4549ae46-dca6-4718-959b-4f02fb98fa59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.734409] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.734753] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d42b943-e7ac-43ff-ae56-18bef3240348 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.759135] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 865.759135] env[65107]: value = "task-5102900" [ 865.759135] env[65107]: _type = "Task" [ 865.759135] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.770407] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102900, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.855063] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "refresh_cache-88cac920-3c1e-41dc-9e52-f51f48a80f56" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.855203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquired lock "refresh_cache-88cac920-3c1e-41dc-9e52-f51f48a80f56" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.855318] env[65107]: DEBUG nova.network.neutron [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 865.977612] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 7b24c530-56bf-4666-96f9-e2dc66728f1a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 866.095519] env[65107]: DEBUG nova.network.neutron [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Successfully created port: d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 866.128487] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 866.277963] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.359569] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.360065] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.433414] env[65107]: DEBUG nova.network.neutron [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 866.448799] env[65107]: DEBUG nova.network.neutron [-] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.477675] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c8fefc-57c1-6114-7073-ba6d5260b95c/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 866.478482] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee79d9b-7132-42ee-9cd9-8286a1b783cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.484654] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 705b91ea-ccd2-41ce-a14f-2ea5b70f80df] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 866.491975] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c8fefc-57c1-6114-7073-ba6d5260b95c/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 866.491975] env[65107]: ERROR oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c8fefc-57c1-6114-7073-ba6d5260b95c/disk-0.vmdk due to incomplete transfer. [ 866.492120] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-02ed1766-9505-4d46-a5b3-dae12b201953 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.501835] env[65107]: DEBUG oslo_vmware.rw_handles [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c8fefc-57c1-6114-7073-ba6d5260b95c/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 866.501882] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Uploaded image 4946aa6f-8b8a-4d0a-955a-55a2bdb46801 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 866.505358] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 866.509253] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0036c5ce-723e-49fe-af76-50305267b649 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.517349] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 866.517349] env[65107]: value = "task-5102901" [ 866.517349] env[65107]: _type = "Task" [ 866.517349] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.524824] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.527583] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.539965] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102901, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.633473] env[65107]: WARNING neutronclient.v2_0.client [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.634250] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.634699] env[65107]: WARNING openstack [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.715074] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f10ce07-eefb-4ce7-afba-2d0ac873814b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.726949] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c341ce-28a4-4cd4-9571-3606151ee152 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.784928] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e70361-e958-4be5-9ae9-50d7d92d6d2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.793179] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102900, 'name': ReconfigVM_Task, 'duration_secs': 0.591848} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.795383] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 369caf24-48a6-4640-8601-e30fd47453da/369caf24-48a6-4640-8601-e30fd47453da.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.796407] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad2c364-392e-42e9-81e8-ea86efe526d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.799656] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488a083d-bace-494e-8066-218fb43c5175 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.805868] env[65107]: DEBUG nova.network.neutron [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Updating instance_info_cache with network_info: [{"id": "2f395da6-1303-4770-b9a5-eed2e9141181", "address": "fa:16:3e:2b:70:ce", "network": {"id": "d625b000-1073-4613-bc0f-033f45135310", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1619844190-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25b948bb3ec94c9886bd2fee74004db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f395da6-13", "ovs_interfaceid": "2f395da6-1303-4770-b9a5-eed2e9141181", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.832489] env[65107]: DEBUG nova.compute.provider_tree [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.838273] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654559f4-e7e2-4918-8104-2b5ae44dcf8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.862651] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babf181c-930f-4ca4-86bc-441f0149f2dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.883805] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e0342f-d589-46a4-bf47-3164c86c0800 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.895299] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.895631] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acc45d38-6037-4f01-a44d-e5cf6014add1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.904638] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 866.904638] env[65107]: value = "task-5102902" [ 866.904638] env[65107]: _type = "Task" [ 866.904638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.913302] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102902, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.953797] env[65107]: INFO nova.compute.manager [-] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Took 1.63 seconds to deallocate network for instance. [ 866.988469] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: e40dc8ad-31cf-4798-ab97-3ee67c8fbe67] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 867.029658] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102901, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.148807] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 867.179693] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 867.179693] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 867.179693] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 867.179693] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 867.179960] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 867.179960] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 867.180772] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.180772] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 867.180772] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 867.181861] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 867.181861] env[65107]: DEBUG nova.virt.hardware [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 867.182187] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da0e809-9947-44b3-97b2-58e363420857 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.192545] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfd44e5-4d2e-4948-a571-d10dce8e629d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.309356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Releasing lock "refresh_cache-88cac920-3c1e-41dc-9e52-f51f48a80f56" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.309735] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Instance network_info: |[{"id": "2f395da6-1303-4770-b9a5-eed2e9141181", "address": "fa:16:3e:2b:70:ce", "network": {"id": "d625b000-1073-4613-bc0f-033f45135310", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1619844190-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25b948bb3ec94c9886bd2fee74004db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f395da6-13", "ovs_interfaceid": "2f395da6-1303-4770-b9a5-eed2e9141181", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 867.310306] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:70:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f92f0b92-d6fb-4d00-8ad5-6b3809ed5493', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f395da6-1303-4770-b9a5-eed2e9141181', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.318935] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Creating folder: Project (25b948bb3ec94c9886bd2fee74004db1). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.318935] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98b426e1-5e32-4682-a803-b31b8ae18565 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.332223] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Created folder: Project (25b948bb3ec94c9886bd2fee74004db1) in parent group-v992574. [ 867.332527] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Creating folder: Instances. Parent ref: group-v992754. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.332816] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cab7e52b-0414-404f-8c34-fe0e669d9e10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.339193] env[65107]: DEBUG nova.scheduler.client.report [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 867.344666] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.347374] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Created folder: Instances in parent group-v992754. [ 867.347685] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 867.348130] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.348298] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3096e5bc-8db5-4e42-b873-36b2ae1c7e61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.371533] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.371533] env[65107]: value = "task-5102905" [ 867.371533] env[65107]: _type = "Task" [ 867.371533] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.382020] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102905, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.417119] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102902, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.462962] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.494737] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 1253473b-d4bc-49ac-9724-9aa1e7a8f038] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 867.529837] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102901, 'name': Destroy_Task, 'duration_secs': 0.804008} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.530523] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Destroyed the VM [ 867.530800] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 867.531208] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4036e033-440b-48ca-b08e-de20b4387eaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.541687] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 867.541687] env[65107]: value = "task-5102906" [ 867.541687] env[65107]: _type = "Task" [ 867.541687] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.554376] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102906, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.790761] env[65107]: DEBUG nova.network.neutron [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Successfully updated port: d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 867.839691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "1a813da6-f777-4888-b2ae-fbad58e01f61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.839963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.847150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.737s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.847748] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 867.851963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 40.899s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.852084] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.854222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.657s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.856525] env[65107]: INFO nova.compute.claims [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.861782] env[65107]: DEBUG nova.compute.manager [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Received event network-changed-2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 867.861782] env[65107]: DEBUG nova.compute.manager [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Refreshing instance network info cache due to event network-changed-2f395da6-1303-4770-b9a5-eed2e9141181. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 867.862026] env[65107]: DEBUG oslo_concurrency.lockutils [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Acquiring lock "refresh_cache-88cac920-3c1e-41dc-9e52-f51f48a80f56" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.862141] env[65107]: DEBUG oslo_concurrency.lockutils [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Acquired lock "refresh_cache-88cac920-3c1e-41dc-9e52-f51f48a80f56" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.862303] env[65107]: DEBUG nova.network.neutron [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Refreshing network info cache for port 2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 867.881103] env[65107]: INFO nova.scheduler.client.report [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted allocations for instance b779b5ad-7d93-47c2-b824-6d76246c00f5 [ 867.889661] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102905, 'name': CreateVM_Task, 'duration_secs': 0.452456} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.890067] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.890567] env[65107]: WARNING neutronclient.v2_0.client [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.890955] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.891125] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.891432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 867.891690] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25a8b499-c310-48cf-9ce8-79d3efeab3e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.897859] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 867.897859] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a9673e-5eb4-0a2a-1fc1-8f11077a4ffe" [ 867.897859] env[65107]: _type = "Task" [ 867.897859] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.907914] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a9673e-5eb4-0a2a-1fc1-8f11077a4ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.917525] env[65107]: DEBUG oslo_vmware.api [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102902, 'name': PowerOnVM_Task, 'duration_secs': 0.521883} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.917799] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.999118] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f58c3088-c821-4eec-be0a-81221debc98e] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 868.054806] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102906, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.295054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.295054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 868.295359] env[65107]: DEBUG nova.network.neutron [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 868.344058] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 868.353945] env[65107]: DEBUG nova.compute.utils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 868.358031] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 868.358031] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 868.358031] env[65107]: WARNING neutronclient.v2_0.client [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.358031] env[65107]: WARNING neutronclient.v2_0.client [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.358031] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.358031] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.367871] env[65107]: WARNING neutronclient.v2_0.client [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.368605] env[65107]: WARNING openstack [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.369079] env[65107]: WARNING openstack [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.393959] env[65107]: DEBUG oslo_concurrency.lockutils [None req-765f0225-67c4-4306-be0d-bbfc5354294f tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "b779b5ad-7d93-47c2-b824-6d76246c00f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.983s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.415798] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a9673e-5eb4-0a2a-1fc1-8f11077a4ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.012289} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.416076] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.416317] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.417031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.417031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 868.417031] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.418698] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2cf0085d-6b8b-4503-b344-9f9019bb2dee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.432037] env[65107]: DEBUG nova.policy [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73470fe64d63489a90389446368d1c51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1ef40d5301a473ea0a1691fecf01fd2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 868.436118] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.436378] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.437457] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be49b113-4f28-4b4a-89c8-91862e5aa9cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.445199] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 868.445199] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4d4db-7b09-34af-290f-64c618c08106" [ 868.445199] env[65107]: _type = "Task" [ 868.445199] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.455677] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4d4db-7b09-34af-290f-64c618c08106, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.505238] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: b7f0fa59-3c6a-4b4b-b400-1f5feec2c0b8] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 868.539305] env[65107]: WARNING openstack [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.539811] env[65107]: WARNING openstack [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.560335] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102906, 'name': RemoveSnapshot_Task, 'duration_secs': 0.591303} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.560335] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 868.560335] env[65107]: DEBUG nova.compute.manager [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 868.561285] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842e1a17-6dab-465a-b35b-45752b0c6f59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.661132] env[65107]: WARNING neutronclient.v2_0.client [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.661898] env[65107]: WARNING openstack [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.662283] env[65107]: WARNING openstack [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.733416] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.733726] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.733954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.734169] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.734355] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.737365] env[65107]: INFO nova.compute.manager [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Terminating instance [ 868.798401] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.799178] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.821388] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Successfully created port: 33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 868.842867] env[65107]: DEBUG nova.network.neutron [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Updated VIF entry in instance network info cache for port 2f395da6-1303-4770-b9a5-eed2e9141181. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 868.843255] env[65107]: DEBUG nova.network.neutron [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Updating instance_info_cache with network_info: [{"id": "2f395da6-1303-4770-b9a5-eed2e9141181", "address": "fa:16:3e:2b:70:ce", "network": {"id": "d625b000-1073-4613-bc0f-033f45135310", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1619844190-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25b948bb3ec94c9886bd2fee74004db1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f92f0b92-d6fb-4d00-8ad5-6b3809ed5493", "external-id": "nsx-vlan-transportzone-48", "segmentation_id": 48, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f395da6-13", "ovs_interfaceid": "2f395da6-1303-4770-b9a5-eed2e9141181", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 868.846715] env[65107]: DEBUG nova.network.neutron [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 868.865987] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 868.871241] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.919402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "e3510c31-d6be-4e9f-a0a2-a662123861e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.919402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.919402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "e3510c31-d6be-4e9f-a0a2-a662123861e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.919402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.919402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.921110] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.921728] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.931908] env[65107]: INFO nova.compute.manager [None req-6db66942-3165-4dd8-910b-fb075a2700e0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance to original state: 'active' [ 868.935357] env[65107]: INFO nova.compute.manager [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Terminating instance [ 868.962024] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4d4db-7b09-34af-290f-64c618c08106, 'name': SearchDatastore_Task, 'duration_secs': 0.017574} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.962024] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a93c2017-85d7-4873-94df-f45b31e1884e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.968137] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 868.968137] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b3714d-ef11-09ac-2740-b8be38067de5" [ 868.968137] env[65107]: _type = "Task" [ 868.968137] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.981786] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b3714d-ef11-09ac-2740-b8be38067de5, 'name': SearchDatastore_Task, 'duration_secs': 0.01074} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.982497] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.982497] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 88cac920-3c1e-41dc-9e52-f51f48a80f56/88cac920-3c1e-41dc-9e52-f51f48a80f56.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.982650] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc191872-27a9-47eb-9dd2-63e42e86ccdb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.001166] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 869.001166] env[65107]: value = "task-5102907" [ 869.001166] env[65107]: _type = "Task" [ 869.001166] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.008873] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.010203] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 0691bdfe-ff23-471b-ad94-2a98adcecd4f] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 869.058409] env[65107]: WARNING neutronclient.v2_0.client [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.058409] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.058409] env[65107]: WARNING openstack [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.078754] env[65107]: INFO nova.compute.manager [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Shelve offloading [ 869.189229] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Successfully created port: c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 869.247568] env[65107]: DEBUG nova.compute.manager [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 869.247568] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-377b393b-b1cf-4fee-bb24-785f19849e44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.260693] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4a3f45-637e-494b-916d-1eec0a204d84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.281115] env[65107]: DEBUG nova.network.neutron [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Updating instance_info_cache with network_info: [{"id": "d5026810-9268-4f3f-9e44-06ea55e41569", "address": "fa:16:3e:b7:67:65", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5026810-92", "ovs_interfaceid": "d5026810-9268-4f3f-9e44-06ea55e41569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 869.310019] env[65107]: WARNING nova.virt.vmwareapi.driver [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc could not be found. [ 869.310019] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.310019] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a0b8410-48c9-4015-9497-f677e4791c0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.321176] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05048278-5daf-4250-89fc-21c7f54221eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.345890] env[65107]: DEBUG oslo_concurrency.lockutils [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] Releasing lock "refresh_cache-88cac920-3c1e-41dc-9e52-f51f48a80f56" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 869.346244] env[65107]: DEBUG nova.compute.manager [req-7532da1f-03c0-4015-be8f-65ffa0765ebe req-bde74ef5-ed75-456f-9e49-cc156eefc3d2 service nova] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Received event network-vif-deleted-a8953e74-c080-4161-885f-b88b84a8bab2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 869.367460] env[65107]: WARNING nova.virt.vmwareapi.vmops [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d65e98bb-f54d-4803-95ca-0fc9b6c184bc could not be found. [ 869.367722] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.367911] env[65107]: INFO nova.compute.manager [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Took 0.12 seconds to destroy the instance on the hypervisor. [ 869.368194] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 869.371593] env[65107]: DEBUG nova.compute.manager [-] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 869.371780] env[65107]: DEBUG nova.network.neutron [-] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 869.371966] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.372554] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.372820] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.444910] env[65107]: DEBUG nova.compute.manager [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 869.445035] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.452606] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efb2aac-0be2-4ae3-b19f-42dfca3a9c47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.464483] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.465017] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d038ceac-0084-4c01-b438-90c7173294a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.474512] env[65107]: DEBUG oslo_vmware.api [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 869.474512] env[65107]: value = "task-5102908" [ 869.474512] env[65107]: _type = "Task" [ 869.474512] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.485444] env[65107]: DEBUG oslo_vmware.api [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.490862] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.513545] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509665} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.514019] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f2814dd6-8be5-49b2-a030-f57e452c2e6d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 869.517428] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 88cac920-3c1e-41dc-9e52-f51f48a80f56/88cac920-3c1e-41dc-9e52-f51f48a80f56.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.517428] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.517428] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d903ebb-aa01-4dac-b9cd-70e275c7e20a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.525274] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 869.525274] env[65107]: value = "task-5102909" [ 869.525274] env[65107]: _type = "Task" [ 869.525274] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.541994] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.583597] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.584349] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b8e8d02-0aec-43f4-9be7-40b9d1ac7d77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.593692] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 869.593692] env[65107]: value = "task-5102910" [ 869.593692] env[65107]: _type = "Task" [ 869.593692] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.596123] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e93bf1f-93d4-4456-9aae-acefb06ee398 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.610947] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948f9bd9-82af-41c0-94ce-3e6838317276 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.615322] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 869.615529] env[65107]: DEBUG nova.compute.manager [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 869.616743] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0e1067-13c3-495e-a77b-e9a8fa826474 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.649313] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbb1b98-f116-4368-ab47-7070dbc926ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.654509] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.654672] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.654851] env[65107]: DEBUG nova.network.neutron [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 869.662283] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a1b11b-1710-4e4b-84c7-757f10bc2d22 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.677956] env[65107]: DEBUG nova.compute.provider_tree [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.783654] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 869.783981] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Instance network_info: |[{"id": "d5026810-9268-4f3f-9e44-06ea55e41569", "address": "fa:16:3e:b7:67:65", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5026810-92", "ovs_interfaceid": "d5026810-9268-4f3f-9e44-06ea55e41569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 869.784469] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:67:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '27138a4c-60c9-45fb-bf37-4c2f765315a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5026810-9268-4f3f-9e44-06ea55e41569', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.792447] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating folder: Project (bfb7fca6907c4c32a1aa6b8023c39afb). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.792715] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-960a04ef-48b1-4412-b58e-b4c2a35d18bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.805092] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created folder: Project (bfb7fca6907c4c32a1aa6b8023c39afb) in parent group-v992574. [ 869.805316] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating folder: Instances. Parent ref: group-v992757. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.805568] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d45a17e-47b7-4cfb-a9f9-6140eb5e21e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.817081] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created folder: Instances in parent group-v992757. [ 869.817354] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 869.817565] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.817783] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e66bed01-2e9d-4d43-bb2b-320eca145539 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.837638] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.837638] env[65107]: value = "task-5102913" [ 869.837638] env[65107]: _type = "Task" [ 869.837638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.847433] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102913, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.886949] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 869.913987] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 869.914252] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 869.914406] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 869.914587] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 869.914732] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 869.914876] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 869.915112] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.915290] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 869.915456] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 869.915625] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 869.915793] env[65107]: DEBUG nova.virt.hardware [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 869.916750] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a192ad-b343-48ae-b2d2-a4139c5b2e1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.925854] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba13568-4d29-47fe-82da-0863cdbbe873 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.992037] env[65107]: DEBUG oslo_vmware.api [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102908, 'name': PowerOffVM_Task, 'duration_secs': 0.261124} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.993525] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.993859] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.994533] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e435b45-d94e-4e57-a5ef-02044b876739 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.019837] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: be518c1d-edd7-40ff-b7cc-3310885b07b4] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 870.037174] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077593} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.037758] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.038672] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a87358-5a9f-42b9-8a19-bdaaee9991c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.073181] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 88cac920-3c1e-41dc-9e52-f51f48a80f56/88cac920-3c1e-41dc-9e52-f51f48a80f56.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.073991] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5a8ba2c-43ad-46b3-b3ea-03c321d5f250 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.100019] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 870.100019] env[65107]: value = "task-5102915" [ 870.100019] env[65107]: _type = "Task" [ 870.100019] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.109750] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.114944] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.115181] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.115446] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleting the datastore file [datastore2] e3510c31-d6be-4e9f-a0a2-a662123861e9 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.115636] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9e962d2-6126-4879-abdd-09f99e874640 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.125796] env[65107]: DEBUG oslo_vmware.api [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for the task: (returnval){ [ 870.125796] env[65107]: value = "task-5102916" [ 870.125796] env[65107]: _type = "Task" [ 870.125796] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.137899] env[65107]: DEBUG oslo_vmware.api [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102916, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.159045] env[65107]: WARNING neutronclient.v2_0.client [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.159371] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.160066] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.183463] env[65107]: DEBUG nova.scheduler.client.report [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 870.357376] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102913, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.376748] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.376924] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.452356] env[65107]: WARNING neutronclient.v2_0.client [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.453033] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.453371] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.524122] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 7a3009bf-54a2-4565-a1aa-1d19286a4810] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 870.526021] env[65107]: DEBUG nova.network.neutron [-] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 870.554751] env[65107]: DEBUG nova.network.neutron [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updating instance_info_cache with network_info: [{"id": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "address": "fa:16:3e:6b:29:71", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bb68bf5-be", "ovs_interfaceid": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 870.611248] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102915, 'name': ReconfigVM_Task, 'duration_secs': 0.337358} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.614397] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 88cac920-3c1e-41dc-9e52-f51f48a80f56/88cac920-3c1e-41dc-9e52-f51f48a80f56.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.614397] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fdac2e62-0251-427c-b775-16134a2d5675 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.623328] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 870.623328] env[65107]: value = "task-5102917" [ 870.623328] env[65107]: _type = "Task" [ 870.623328] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.633333] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102917, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.639024] env[65107]: DEBUG oslo_vmware.api [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Task: {'id': task-5102916, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167809} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.639376] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.639613] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.640228] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.640228] env[65107]: INFO nova.compute.manager [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Took 1.20 seconds to destroy the instance on the hypervisor. [ 870.640481] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 870.640600] env[65107]: DEBUG nova.compute.manager [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 870.640743] env[65107]: DEBUG nova.network.neutron [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 870.641067] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.641650] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.641977] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.689905] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.836s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.690467] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 870.696028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.783s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.696028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.696028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 36.418s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.696028] env[65107]: DEBUG nova.objects.instance [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 870.714225] env[65107]: DEBUG nova.compute.manager [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Received event network-vif-plugged-d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 870.714316] env[65107]: DEBUG oslo_concurrency.lockutils [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Acquiring lock "ee50d08a-57fc-4c05-96fe-a11fe2708165-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.715025] env[65107]: DEBUG oslo_concurrency.lockutils [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.715025] env[65107]: DEBUG oslo_concurrency.lockutils [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.715025] env[65107]: DEBUG nova.compute.manager [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] No waiting events found dispatching network-vif-plugged-d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 870.715025] env[65107]: WARNING nova.compute.manager [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Received unexpected event network-vif-plugged-d5026810-9268-4f3f-9e44-06ea55e41569 for instance with vm_state building and task_state spawning. [ 870.715394] env[65107]: DEBUG nova.compute.manager [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Received event network-changed-d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 870.715560] env[65107]: DEBUG nova.compute.manager [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Refreshing instance network info cache due to event network-changed-d5026810-9268-4f3f-9e44-06ea55e41569. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 870.715743] env[65107]: DEBUG oslo_concurrency.lockutils [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Acquiring lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.715980] env[65107]: DEBUG oslo_concurrency.lockutils [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Acquired lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.716063] env[65107]: DEBUG nova.network.neutron [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Refreshing network info cache for port d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 870.719138] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.737952] env[65107]: INFO nova.scheduler.client.report [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Deleted allocations for instance ef417f32-8c0e-4a62-9078-cb2bf95b85f8 [ 870.848782] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102913, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.854206] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Successfully updated port: 33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 871.027837] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: eac4b8f9-9cd5-44a3-a12e-9ec22767b907] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 871.031829] env[65107]: INFO nova.compute.manager [-] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Took 1.66 seconds to deallocate network for instance. [ 871.059831] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 871.060162] env[65107]: WARNING neutronclient.v2_0.client [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.060811] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.061781] env[65107]: WARNING openstack [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.066967] env[65107]: WARNING neutronclient.v2_0.client [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.132877] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102917, 'name': Rename_Task, 'duration_secs': 0.154381} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.133167] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.133418] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce0e819d-b4bd-4d6d-adf4-ed9d638040de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.141659] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 871.141659] env[65107]: value = "task-5102918" [ 871.141659] env[65107]: _type = "Task" [ 871.141659] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.155638] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.209822] env[65107]: DEBUG nova.compute.utils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 871.210766] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 871.211014] env[65107]: DEBUG nova.network.neutron [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 871.211344] env[65107]: WARNING neutronclient.v2_0.client [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.211634] env[65107]: WARNING neutronclient.v2_0.client [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.212240] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.212658] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.221079] env[65107]: WARNING neutronclient.v2_0.client [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.221736] env[65107]: WARNING openstack [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.222359] env[65107]: WARNING openstack [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.253082] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52f6a7a0-1ee7-4135-ac73-ea32afcbaed1 tempest-FloatingIPsAssociationTestJSON-1065123087 tempest-FloatingIPsAssociationTestJSON-1065123087-project-member] Lock "ef417f32-8c0e-4a62-9078-cb2bf95b85f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.888s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.282234] env[65107]: DEBUG nova.policy [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61d0444710414b8eae5e40ed898d8712', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb7fca6907c4c32a1aa6b8023c39afb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 871.351551] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102913, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.415715] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.416671] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd040dd-64d6-46a5-b9ac-c0c4858ec683 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.425194] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 871.426538] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f10946f-98c8-417a-8e97-4ac4e9095147 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.429692] env[65107]: DEBUG nova.compute.manager [req-b716fd57-81e7-40c9-a4f0-a20d4399ff23 req-d673f70d-2b7d-4189-a18b-5a2c9bfe45ad service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received event network-vif-plugged-33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 871.429927] env[65107]: DEBUG oslo_concurrency.lockutils [req-b716fd57-81e7-40c9-a4f0-a20d4399ff23 req-d673f70d-2b7d-4189-a18b-5a2c9bfe45ad service nova] Acquiring lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.430163] env[65107]: DEBUG oslo_concurrency.lockutils [req-b716fd57-81e7-40c9-a4f0-a20d4399ff23 req-d673f70d-2b7d-4189-a18b-5a2c9bfe45ad service nova] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.430332] env[65107]: DEBUG oslo_concurrency.lockutils [req-b716fd57-81e7-40c9-a4f0-a20d4399ff23 req-d673f70d-2b7d-4189-a18b-5a2c9bfe45ad service nova] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.430523] env[65107]: DEBUG nova.compute.manager [req-b716fd57-81e7-40c9-a4f0-a20d4399ff23 req-d673f70d-2b7d-4189-a18b-5a2c9bfe45ad service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] No waiting events found dispatching network-vif-plugged-33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 871.430656] env[65107]: WARNING nova.compute.manager [req-b716fd57-81e7-40c9-a4f0-a20d4399ff23 req-d673f70d-2b7d-4189-a18b-5a2c9bfe45ad service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received unexpected event network-vif-plugged-33f13ff6-7d02-4931-bda5-04c0658cfd6c for instance with vm_state building and task_state spawning. [ 871.497944] env[65107]: DEBUG nova.network.neutron [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 871.508804] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 871.509053] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 871.509249] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleting the datastore file [datastore2] 701a1a16-9348-4748-bb24-ee245a76566c {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.509536] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9949502-e12a-4bef-a089-2a28bee6c339 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.518290] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 871.518290] env[65107]: value = "task-5102920" [ 871.518290] env[65107]: _type = "Task" [ 871.518290] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.530834] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.531374] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: abb9dc0e-ce18-49c8-a472-d39a3fcec887] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 871.585296] env[65107]: INFO nova.compute.manager [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Took 0.55 seconds to detach 1 volumes for instance. [ 871.590697] env[65107]: DEBUG nova.compute.manager [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Deleting volume: f04439cd-6130-4943-97ee-01d8501c2efc {{(pid=65107) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 871.659501] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102918, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.694379] env[65107]: DEBUG nova.network.neutron [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Successfully created port: 4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 871.713138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3ce65c47-7ceb-4ce2-9adf-f45e0f6ac118 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.714437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.399s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.716685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.717050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.106s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.717385] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.719242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.338s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.719459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.721923] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.187s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.723754] env[65107]: INFO nova.compute.claims [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.728795] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 871.768249] env[65107]: INFO nova.scheduler.client.report [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Deleted allocations for instance 5fccc802-f5ac-4450-8246-4cf9a5371046 [ 871.773266] env[65107]: INFO nova.scheduler.client.report [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted allocations for instance d2d0e18c-4839-416c-b363-b93996e8ba7f [ 871.789931] env[65107]: INFO nova.scheduler.client.report [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Deleted allocations for instance 780f910f-4f73-41ba-a795-0daae3097314 [ 871.858298] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102913, 'name': CreateVM_Task, 'duration_secs': 1.520387} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.858925] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.859674] env[65107]: WARNING neutronclient.v2_0.client [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.860198] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.860402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.861076] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 871.862023] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09fdba0f-e416-4e81-a5c9-43a8d04bd3ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.868314] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 871.868314] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5205eda9-d4fe-ce01-a66b-7f60c5dd3548" [ 871.868314] env[65107]: _type = "Task" [ 871.868314] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.879450] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5205eda9-d4fe-ce01-a66b-7f60c5dd3548, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.923710] env[65107]: WARNING openstack [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.924152] env[65107]: WARNING openstack [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.000517] env[65107]: INFO nova.compute.manager [-] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Took 1.36 seconds to deallocate network for instance. [ 872.031505] env[65107]: DEBUG oslo_vmware.api [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5102920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186504} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.031806] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.032115] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 872.032535] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.035906] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: ccd68268-d0fc-406e-896f-c61b114dd75d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 872.060639] env[65107]: INFO nova.scheduler.client.report [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocations for instance 701a1a16-9348-4748-bb24-ee245a76566c [ 872.132952] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.153693] env[65107]: DEBUG oslo_vmware.api [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5102918, 'name': PowerOnVM_Task, 'duration_secs': 0.675684} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.153966] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.154186] env[65107]: INFO nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Took 7.78 seconds to spawn the instance on the hypervisor. [ 872.154357] env[65107]: DEBUG nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 872.155120] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c6eaff-0cc6-49d0-aa7f-8cf967b56f15 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.289691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39fc1e09-50b6-471b-acb8-0d43fe57cb6c tempest-ServerTagsTestJSON-1705843489 tempest-ServerTagsTestJSON-1705843489-project-member] Lock "5fccc802-f5ac-4450-8246-4cf9a5371046" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.448s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.290896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bd357d47-42f2-4989-bf64-d8e269623652 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "d2d0e18c-4839-416c-b363-b93996e8ba7f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.412s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.299059] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8fcc6e10-cf8a-4c60-a725-527f69598937 tempest-ServerShowV254Test-1225384254 tempest-ServerShowV254Test-1225384254-project-member] Lock "780f910f-4f73-41ba-a795-0daae3097314" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.535s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.381217] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5205eda9-d4fe-ce01-a66b-7f60c5dd3548, 'name': SearchDatastore_Task, 'duration_secs': 0.015398} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.381608] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.381895] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.382188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.382362] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 872.382611] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.382900] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-046ff866-118b-4319-8fd3-9bb1cc92a788 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.397507] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.397684] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.398424] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2186ad3-e98c-4be1-90ec-b62b18843c67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.404587] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 872.404587] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52418428-ce43-cdac-16e9-830bead49979" [ 872.404587] env[65107]: _type = "Task" [ 872.404587] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.413399] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52418428-ce43-cdac-16e9-830bead49979, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.509020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.538972] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 0ef89aea-5373-4d07-9587-37890c08d35b] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 872.567243] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.673866] env[65107]: INFO nova.compute.manager [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Took 57.11 seconds to build instance. [ 872.739323] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 872.772656] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 872.772934] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 872.773105] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 872.773288] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 872.773429] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 872.773571] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 872.773778] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.773950] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 872.774126] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 872.774301] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 872.774471] env[65107]: DEBUG nova.virt.hardware [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 872.775382] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789a47ba-9119-425c-bba8-0d29144758b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.786777] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc53a71-a3ab-4833-a560-7b10719c2b52 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.915982] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52418428-ce43-cdac-16e9-830bead49979, 'name': SearchDatastore_Task, 'duration_secs': 0.040501} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.919773] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee87d6e9-1dbf-418d-bfef-697b347d8098 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.926867] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 872.926867] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fe6ccd-baa5-5164-2e2f-738176579603" [ 872.926867] env[65107]: _type = "Task" [ 872.926867] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.938983] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fe6ccd-baa5-5164-2e2f-738176579603, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.032744] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Successfully updated port: c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 873.041769] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 4e70aaf3-5f1c-4a61-a790-a0a5e409170d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 873.122982] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d532c40a-42c7-416c-9b72-150bfccb9b59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.131441] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd9552b-ae85-4c15-90c6-e5e5ee951528 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.165270] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41034d6-9718-48a9-917e-b3f213e66277 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.175095] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb51d12-42bd-4f02-837d-7d5860f672ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.179750] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d8b61c-4407-46d9-9000-2cd6363bd607 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.626s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.191259] env[65107]: DEBUG nova.compute.provider_tree [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.438907] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fe6ccd-baa5-5164-2e2f-738176579603, 'name': SearchDatastore_Task, 'duration_secs': 0.040894} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.439221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 873.439486] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/ee50d08a-57fc-4c05-96fe-a11fe2708165.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.439754] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e749df2f-9514-4e41-b9a0-0caf383875ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.448059] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 873.448059] env[65107]: value = "task-5102922" [ 873.448059] env[65107]: _type = "Task" [ 873.448059] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.457459] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.510343] env[65107]: DEBUG nova.network.neutron [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Successfully updated port: 4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 873.535405] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.535607] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.535849] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 873.545998] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 31195587-86ce-4379-946d-1abb51c35f08] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 873.695677] env[65107]: DEBUG nova.scheduler.client.report [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.959271] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502618} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.959502] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/ee50d08a-57fc-4c05-96fe-a11fe2708165.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.959731] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.959946] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce83c23e-e713-4c82-a838-c9f95184f0d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.966328] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 873.966328] env[65107]: value = "task-5102923" [ 873.966328] env[65107]: _type = "Task" [ 873.966328] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.974875] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.013498] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.013615] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.013814] env[65107]: DEBUG nova.network.neutron [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 874.039216] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.039660] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.049467] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: ba2ed98f-ec1e-4a98-a169-7b2a1fbcb4e1] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 874.077702] env[65107]: WARNING neutronclient.v2_0.client [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.078401] env[65107]: WARNING openstack [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.078739] env[65107]: WARNING openstack [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.201683] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.202126] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 874.204989] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.403s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.206457] env[65107]: INFO nova.compute.claims [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.480835] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068792} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.481059] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.482448] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764fbfe3-a99a-427e-9611-44903b810939 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.509231] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/ee50d08a-57fc-4c05-96fe-a11fe2708165.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.509615] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7930ee07-fc56-4b8e-90d2-07acb2906429 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.528590] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.529016] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.544519] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 874.544519] env[65107]: value = "task-5102924" [ 874.544519] env[65107]: _type = "Task" [ 874.544519] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.550694] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 874.554279] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 874.554336] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Cleaning up deleted instances with incomplete migration {{(pid=65107) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 874.560726] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102924, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.712568] env[65107]: DEBUG nova.compute.utils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 874.716689] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 874.720554] env[65107]: DEBUG nova.network.neutron [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 874.723420] env[65107]: WARNING neutronclient.v2_0.client [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.723420] env[65107]: WARNING neutronclient.v2_0.client [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.724317] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.726206] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.057155] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102924, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.057445] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 875.080663] env[65107]: DEBUG nova.network.neutron [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 875.102484] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.102873] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.200308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955aea6d-b5dc-49dd-96ec-a96eb932b9d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.211263] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5980d934-2d49-4d16-ba80-2efa477e8f5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.217618] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 875.253371] env[65107]: DEBUG nova.network.neutron [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Updated VIF entry in instance network info cache for port d5026810-9268-4f3f-9e44-06ea55e41569. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 875.254475] env[65107]: DEBUG nova.network.neutron [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Updating instance_info_cache with network_info: [{"id": "d5026810-9268-4f3f-9e44-06ea55e41569", "address": "fa:16:3e:b7:67:65", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5026810-92", "ovs_interfaceid": "d5026810-9268-4f3f-9e44-06ea55e41569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 875.255646] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a74185-b805-40ce-8fb2-d3f5b73e33c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.265869] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbab20f-eb1f-415a-9faa-4f8c176c225e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.281884] env[65107]: DEBUG nova.compute.provider_tree [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.286688] env[65107]: DEBUG nova.policy [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd87f09a43c049deb1fda21602b55a64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '544c649453844164bac53c98b342543f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 875.398334] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.398723] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.556841] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102924, 'name': ReconfigVM_Task, 'duration_secs': 0.522305} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.557161] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Reconfigured VM instance instance-0000003d to attach disk [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/ee50d08a-57fc-4c05-96fe-a11fe2708165.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.557823] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ae62df8-4a1c-448d-88a1-e36203dfd489 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.576532] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 875.576532] env[65107]: value = "task-5102925" [ 875.576532] env[65107]: _type = "Task" [ 875.576532] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.590017] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102925, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.759278] env[65107]: DEBUG oslo_concurrency.lockutils [req-fc67b785-ba71-4013-820b-becf1105714d req-fa8a72fc-0762-4298-ab04-677994d1927d service nova] Releasing lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.791175] env[65107]: DEBUG nova.scheduler.client.report [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 875.869595] env[65107]: WARNING neutronclient.v2_0.client [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.870409] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.870949] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.981726] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "369caf24-48a6-4640-8601-e30fd47453da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.982090] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.982307] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "369caf24-48a6-4640-8601-e30fd47453da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.982539] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.982792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 875.985596] env[65107]: INFO nova.compute.manager [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Terminating instance [ 876.086915] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102925, 'name': Rename_Task, 'duration_secs': 0.160172} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.087273] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.087608] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fc3054a-916c-4f5e-afae-8afc4377a052 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.095275] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 876.095275] env[65107]: value = "task-5102926" [ 876.095275] env[65107]: _type = "Task" [ 876.095275] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.106327] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102926, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.200671] env[65107]: WARNING neutronclient.v2_0.client [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.201635] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.201990] env[65107]: WARNING openstack [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.261567] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 876.290929] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 876.291249] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.291482] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 876.291585] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.291728] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 876.292018] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 876.292159] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.292338] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 876.292507] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 876.292667] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 876.292837] env[65107]: DEBUG nova.virt.hardware [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 876.293912] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b266f29b-c212-44d0-8436-fb598c520a71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.298211] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.093s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.298726] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 876.301496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.726s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.301735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.305472] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.376s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.305472] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.308744] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.112s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.310777] env[65107]: INFO nova.compute.claims [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.321505] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce412fae-6267-4c3b-9c47-4967f8b3f316 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.350824] env[65107]: INFO nova.scheduler.client.report [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Deleted allocations for instance 7062b119-01a4-45f4-bb52-212178382271 [ 876.357912] env[65107]: INFO nova.scheduler.client.report [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Deleted allocations for instance 805e2210-fb58-48da-be63-5d548b2f8a2f [ 876.368726] env[65107]: DEBUG nova.network.neutron [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Successfully created port: 875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 876.489844] env[65107]: DEBUG nova.compute.manager [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 876.490195] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.491604] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cadb556-e940-4102-be2a-16be30402461 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.501195] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.501486] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33942736-2e54-43c7-bbb2-fc299c23708a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.508479] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 876.508479] env[65107]: value = "task-5102927" [ 876.508479] env[65107]: _type = "Task" [ 876.508479] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.518537] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.610210] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102926, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.817012] env[65107]: DEBUG nova.compute.utils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 876.828705] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 876.828705] env[65107]: DEBUG nova.network.neutron [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 876.828705] env[65107]: WARNING neutronclient.v2_0.client [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.828705] env[65107]: WARNING neutronclient.v2_0.client [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.828705] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.828705] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.863768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5ed87da-277a-43cc-a3bb-344d97c6df5c tempest-ServersTestBootFromVolume-630499006 tempest-ServersTestBootFromVolume-630499006-project-member] Lock "7062b119-01a4-45f4-bb52-212178382271" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.018s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.868926] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2dbb7c12-0b61-4a93-af22-bfe10b006901 tempest-VolumesAdminNegativeTest-870844795 tempest-VolumesAdminNegativeTest-870844795-project-member] Lock "805e2210-fb58-48da-be63-5d548b2f8a2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.806s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.019549] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102927, 'name': PowerOffVM_Task, 'duration_secs': 0.339904} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.022546] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.022764] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.023622] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87f75f3b-5dc9-4833-8a7f-e60c423e3f48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.098024] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.098024] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.098024] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleting the datastore file [datastore1] 369caf24-48a6-4640-8601-e30fd47453da {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.098024] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49b8731a-a290-43d6-822e-b8912e86dbe1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.112417] env[65107]: DEBUG oslo_vmware.api [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102926, 'name': PowerOnVM_Task, 'duration_secs': 0.820379} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.113912] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.114149] env[65107]: INFO nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Took 9.97 seconds to spawn the instance on the hypervisor. [ 877.114337] env[65107]: DEBUG nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 877.114676] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 877.114676] env[65107]: value = "task-5102929" [ 877.114676] env[65107]: _type = "Task" [ 877.114676] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.115535] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c3217e-1642-4b8f-8ad4-e623889203ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.131586] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.241461] env[65107]: DEBUG nova.network.neutron [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Updating instance_info_cache with network_info: [{"id": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "address": "fa:16:3e:d4:a3:9f", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4659264c-96", "ovs_interfaceid": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 877.254420] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.254916] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.266211] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7ff6ca-4329-40b7-94c2-f7d6cf24a3aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.277030] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4a86d7-1c40-4412-adcc-3a2bdf10ce1e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.317188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c0e673-c1a0-4760-bb85-7e458868d092 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.324295] env[65107]: DEBUG nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Received event network-vif-deleted-33799b3d-5ade-4cb0-ae83-3d00bfd972ca {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 877.324692] env[65107]: DEBUG nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Received event network-vif-deleted-2686e3e0-c528-42bb-9444-050209faa6b1 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 877.324975] env[65107]: DEBUG nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Received event network-vif-unplugged-8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 877.325503] env[65107]: DEBUG oslo_concurrency.lockutils [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Acquiring lock "701a1a16-9348-4748-bb24-ee245a76566c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.325625] env[65107]: DEBUG oslo_concurrency.lockutils [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Lock "701a1a16-9348-4748-bb24-ee245a76566c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.325940] env[65107]: DEBUG oslo_concurrency.lockutils [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Lock "701a1a16-9348-4748-bb24-ee245a76566c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.326244] env[65107]: DEBUG nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] No waiting events found dispatching network-vif-unplugged-8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 877.326533] env[65107]: WARNING nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Received unexpected event network-vif-unplugged-8bb68bf5-be11-4086-a291-e8d4eeb2047b for instance with vm_state shelved_offloaded and task_state None. [ 877.326831] env[65107]: DEBUG nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Received event network-changed-8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 877.327204] env[65107]: DEBUG nova.compute.manager [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Refreshing instance network info cache due to event network-changed-8bb68bf5-be11-4086-a291-e8d4eeb2047b. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 877.327479] env[65107]: DEBUG oslo_concurrency.lockutils [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Acquiring lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.327664] env[65107]: DEBUG oslo_concurrency.lockutils [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Acquired lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.327941] env[65107]: DEBUG nova.network.neutron [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Refreshing network info cache for port 8bb68bf5-be11-4086-a291-e8d4eeb2047b {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 877.333217] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 877.342025] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034f7f8b-d2ad-4946-9f23-cc991c30edd0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.359431] env[65107]: DEBUG nova.compute.provider_tree [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.365309] env[65107]: DEBUG nova.policy [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c5b4a751c1e4a3086f7e3359042582b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b981c5f6b3b64476a2406cd9549f7d6b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 877.629185] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.647055] env[65107]: INFO nova.compute.manager [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Took 51.83 seconds to build instance. [ 877.744292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.744771] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Instance network_info: |[{"id": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "address": "fa:16:3e:d4:a3:9f", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4659264c-96", "ovs_interfaceid": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 877.745157] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:a3:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '27138a4c-60c9-45fb-bf37-4c2f765315a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4659264c-9645-4cc9-a6aa-4d3299ae4c31', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.753135] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 877.753135] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.753360] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e1fa3cf-8148-4ae5-b930-eb622f978bdc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.777308] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.777308] env[65107]: value = "task-5102930" [ 877.777308] env[65107]: _type = "Task" [ 877.777308] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.787352] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102930, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.831679] env[65107]: WARNING neutronclient.v2_0.client [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.832420] env[65107]: WARNING openstack [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.833238] env[65107]: WARNING openstack [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.868842] env[65107]: DEBUG nova.scheduler.client.report [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 878.130076] env[65107]: DEBUG oslo_vmware.api [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5102929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.61514} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.130360] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.130753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.130753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.130890] env[65107]: INFO nova.compute.manager [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Took 1.64 seconds to destroy the instance on the hypervisor. [ 878.131298] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 878.131393] env[65107]: DEBUG nova.compute.manager [-] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 878.131493] env[65107]: DEBUG nova.network.neutron [-] [instance: 369caf24-48a6-4640-8601-e30fd47453da] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 878.131817] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.132415] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.132678] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.149285] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2d1603da-3019-4170-aa72-771f9762f4b0 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.342s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.251067] env[65107]: DEBUG nova.network.neutron [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Successfully updated port: 875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 878.290468] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102930, 'name': CreateVM_Task, 'duration_secs': 0.48137} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.290643] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.291262] env[65107]: WARNING neutronclient.v2_0.client [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.291717] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.291918] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.292396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 878.292709] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a37b2dac-c3f5-4615-8b35-3941e9e5ba0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.298890] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 878.298890] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bdffec-1da0-59d7-75f5-df55057a03fa" [ 878.298890] env[65107]: _type = "Task" [ 878.298890] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.308748] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bdffec-1da0-59d7-75f5-df55057a03fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.347015] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 878.377235] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 878.377235] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 878.377235] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 878.377235] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 878.377470] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 878.377512] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 878.377740] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.377904] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 878.378173] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 878.378354] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 878.379685] env[65107]: DEBUG nova.virt.hardware [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 878.379685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.073s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.381935] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddc5284-8c95-4590-bec4-171925414965 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.385097] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.304s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.393213] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad5cf1b-6b09-4010-bef5-63b77233edac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.411210] env[65107]: DEBUG nova.network.neutron [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Successfully created port: 190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 878.458452] env[65107]: WARNING neutronclient.v2_0.client [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.459162] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.459530] env[65107]: WARNING openstack [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.475071] env[65107]: DEBUG nova.compute.manager [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received event network-changed-33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 878.475373] env[65107]: DEBUG nova.compute.manager [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Refreshing instance network info cache due to event network-changed-33f13ff6-7d02-4931-bda5-04c0658cfd6c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 878.475573] env[65107]: DEBUG oslo_concurrency.lockutils [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Acquiring lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.717854] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.759113] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.759335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.759531] env[65107]: DEBUG nova.network.neutron [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 878.817823] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bdffec-1da0-59d7-75f5-df55057a03fa, 'name': SearchDatastore_Task, 'duration_secs': 0.020311} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.818895] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.818895] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.819158] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.820065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.820065] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.820065] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4da16d02-31d7-4f84-9e3a-b69d0525af4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.834611] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.834611] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.834611] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05288a9e-7605-452a-b9a0-717af1db722e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.843490] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 878.843490] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cb833-8ce7-1be4-3357-a59a40a10dc8" [ 878.843490] env[65107]: _type = "Task" [ 878.843490] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.854300] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cb833-8ce7-1be4-3357-a59a40a10dc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.886609] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "dc0e0c48-6f6e-4e64-a30d-1dfa99213365" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.887554] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "dc0e0c48-6f6e-4e64-a30d-1dfa99213365" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.898617] env[65107]: INFO nova.compute.claims [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.265570] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.265988] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.360147] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cb833-8ce7-1be4-3357-a59a40a10dc8, 'name': SearchDatastore_Task, 'duration_secs': 0.020599} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.361473] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-373f693d-ea75-4919-91cc-c963a90d5d39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.369678] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 879.369678] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527ea6a1-775b-2df6-d2bd-e2db56574130" [ 879.369678] env[65107]: _type = "Task" [ 879.369678] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.375259] env[65107]: DEBUG oslo_concurrency.lockutils [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "701a1a16-9348-4748-bb24-ee245a76566c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.384798] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527ea6a1-775b-2df6-d2bd-e2db56574130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.393155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "dc0e0c48-6f6e-4e64-a30d-1dfa99213365" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.506s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.393766] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 879.403767] env[65107]: INFO nova.compute.resource_tracker [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating resource usage from migration cb52d55b-9ce2-4ccf-996d-1ceb56e3a62b [ 879.783290] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207b9230-42dd-45f1-8ecd-6a7ea706510e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.793437] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef623c83-e2dd-465a-9e51-76d90a069272 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.831305] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e157d51d-ff3a-497e-a77c-8b2b6219f936 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.840678] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21177521-1ea0-4d46-ad51-75b1ab9eb3eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.858885] env[65107]: DEBUG nova.compute.provider_tree [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.882488] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527ea6a1-775b-2df6-d2bd-e2db56574130, 'name': SearchDatastore_Task, 'duration_secs': 0.013073} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.882808] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.883079] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/40e17da0-b459-4619-8e9e-18c3093328c4.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.883401] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-338bec5a-a153-4229-a36c-1d31ae6291e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.892454] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 879.892454] env[65107]: value = "task-5102931" [ 879.892454] env[65107]: _type = "Task" [ 879.892454] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.899379] env[65107]: DEBUG nova.compute.utils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 879.904135] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 879.904368] env[65107]: DEBUG nova.network.neutron [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 879.904698] env[65107]: WARNING neutronclient.v2_0.client [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.904993] env[65107]: WARNING neutronclient.v2_0.client [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.905965] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.905965] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.913261] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.352659] env[65107]: DEBUG nova.network.neutron [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Successfully updated port: 190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 880.363643] env[65107]: DEBUG nova.scheduler.client.report [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 880.412848] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102931, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.414943] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 880.544854] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.545040] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.692020] env[65107]: DEBUG nova.network.neutron [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 880.857724] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.857724] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.857724] env[65107]: DEBUG nova.network.neutron [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 880.871371] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.484s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.871371] env[65107]: INFO nova.compute.manager [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Migrating [ 880.890342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.458s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.893406] env[65107]: INFO nova.compute.claims [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.916990] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102931, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.064058] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.064605] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.064605] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.064605] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.064732] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.065130] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.065130] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 881.065543] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.147154] env[65107]: DEBUG nova.network.neutron [-] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.272943] env[65107]: DEBUG nova.network.neutron [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Updating instance_info_cache with network_info: [{"id": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "address": "fa:16:3e:85:cd:c3", "network": {"id": "cf8b5cec-0dbb-48c0-8344-638b7845cd5b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1028060705", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f13ff6-7d", "ovs_interfaceid": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "address": "fa:16:3e:66:c6:76", "network": {"id": "31093185-4798-44f4-a16c-a1b6b13e74ce", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-366736123", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fdc887-a8", "ovs_interfaceid": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.292925] env[65107]: WARNING openstack [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.293422] env[65107]: WARNING openstack [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.363548] env[65107]: DEBUG nova.policy [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0201c5835366414b8557dd369285464f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79f6cf60cda141238e1b8ae1a1a32c22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 881.366789] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.367848] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.406691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.406865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.407036] env[65107]: DEBUG nova.network.neutron [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 881.419181] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102931, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.424612] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 881.451879] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 881.452182] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 881.452439] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 881.452757] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 881.452859] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 881.453026] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 881.453295] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.453490] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 881.453686] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 881.453878] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 881.454093] env[65107]: DEBUG nova.virt.hardware [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 881.455310] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117a86f4-4fc8-41a8-b115-60813a293d35 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.464316] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d24b843-f7c5-454d-949d-7ada2591964f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.569551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.589473] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.590157] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.650080] env[65107]: INFO nova.compute.manager [-] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Took 3.52 seconds to deallocate network for instance. [ 881.777871] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Releasing lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.778331] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Instance network_info: |[{"id": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "address": "fa:16:3e:85:cd:c3", "network": {"id": "cf8b5cec-0dbb-48c0-8344-638b7845cd5b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1028060705", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f13ff6-7d", "ovs_interfaceid": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "address": "fa:16:3e:66:c6:76", "network": {"id": "31093185-4798-44f4-a16c-a1b6b13e74ce", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-366736123", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fdc887-a8", "ovs_interfaceid": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 881.778700] env[65107]: DEBUG oslo_concurrency.lockutils [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Acquired lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.778867] env[65107]: DEBUG nova.network.neutron [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Refreshing network info cache for port 33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 881.782053] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:cd:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b107fab-ee71-47db-ad4d-3c6f05546843', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33f13ff6-7d02-4931-bda5-04c0658cfd6c', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:c6:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '669e4919-e0ad-4e23-9f23-4c5f2be0d858', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2fdc887-a837-4102-9f3a-85fec34ad2b9', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.792884] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 881.794150] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.794961] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5073c232-f0c9-4d2d-ac94-0b044d29e493 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.821637] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.821637] env[65107]: value = "task-5102932" [ 881.821637] env[65107]: _type = "Task" [ 881.821637] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.831304] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102932, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.922049] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.926025] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.926025] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.947302] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102931, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.658987} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.950876] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/40e17da0-b459-4619-8e9e-18c3093328c4.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.950876] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.951513] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42a8ba37-f720-4026-b85f-3193e75ba5d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.961066] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 881.961066] env[65107]: value = "task-5102933" [ 881.961066] env[65107]: _type = "Task" [ 881.961066] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.974587] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102933, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.157981] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.235526] env[65107]: DEBUG nova.network.neutron [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 882.296061] env[65107]: WARNING neutronclient.v2_0.client [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.296641] env[65107]: WARNING openstack [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.297745] env[65107]: WARNING openstack [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.339040] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102932, 'name': CreateVM_Task, 'duration_secs': 0.478435} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.339156] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.339698] env[65107]: WARNING neutronclient.v2_0.client [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.340496] env[65107]: WARNING neutronclient.v2_0.client [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.340496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.340726] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 882.341043] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 882.341239] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5f9bed4-2f5f-4eb7-93cf-fd26db5a184f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.351468] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 882.351468] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52587c6d-40f9-200a-d1a1-aa4acbffefea" [ 882.351468] env[65107]: _type = "Task" [ 882.351468] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.362975] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52587c6d-40f9-200a-d1a1-aa4acbffefea, 'name': SearchDatastore_Task, 'duration_secs': 0.012244} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.363465] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.363814] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.364388] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.364986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 882.366122] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.368595] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b79e8b9f-7497-4581-824a-902fd6d2aa10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.385024] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.385024] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.385024] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5699dfa6-a210-40c0-b721-de900d5ee5e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.392138] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 882.392138] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5200b216-ebc9-b6e4-e487-ee23fce0f5f2" [ 882.392138] env[65107]: _type = "Task" [ 882.392138] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.407530] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5200b216-ebc9-b6e4-e487-ee23fce0f5f2, 'name': SearchDatastore_Task, 'duration_secs': 0.010973} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.408105] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d996b33-5332-4585-be21-1f2c2301625f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.413133] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4327bd-edfd-4867-8b80-1d77e701a9cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.417481] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 882.417481] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e8ea58-1f20-2d21-8f3d-ffec2dda3009" [ 882.417481] env[65107]: _type = "Task" [ 882.417481] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.425022] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c07f9f-4e95-460d-bc72-0d05a0266c40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.432296] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e8ea58-1f20-2d21-8f3d-ffec2dda3009, 'name': SearchDatastore_Task, 'duration_secs': 0.011229} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.432966] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.433392] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e/64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.433537] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6480687f-4384-4291-9a3f-1a2d6dd2832c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.474583] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66e0a9e-0b0a-4895-b9e5-c04f23a3fc46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.478668] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 882.478668] env[65107]: value = "task-5102934" [ 882.478668] env[65107]: _type = "Task" [ 882.478668] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.489123] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102933, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093567} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.490097] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.491249] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b429cbca-c7ee-49e8-a90b-c4ce14f0d9a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.499793] env[65107]: WARNING neutronclient.v2_0.client [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.500531] env[65107]: WARNING openstack [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.501152] env[65107]: WARNING openstack [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.510482] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56291cb-9a6a-4328-9336-f50c343d35e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.512829] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.536890] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/40e17da0-b459-4619-8e9e-18c3093328c4.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.545882] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f59f8ed6-40fc-41bc-842a-3854efccbd65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.560924] env[65107]: DEBUG nova.compute.provider_tree [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.563733] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.564651] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.577408] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 882.577408] env[65107]: value = "task-5102935" [ 882.577408] env[65107]: _type = "Task" [ 882.577408] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.588711] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.599656] env[65107]: WARNING neutronclient.v2_0.client [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.602024] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.602024] env[65107]: WARNING openstack [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.732766] env[65107]: DEBUG nova.network.neutron [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Successfully created port: c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 883.004329] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102934, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.073822] env[65107]: DEBUG nova.scheduler.client.report [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 883.085562] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.086417] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.102610] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102935, 'name': ReconfigVM_Task, 'duration_secs': 0.475301} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.102915] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/40e17da0-b459-4619-8e9e-18c3093328c4.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.104527] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2700e208-b4e4-4280-9235-a60c1511567b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.114379] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 883.114379] env[65107]: value = "task-5102936" [ 883.114379] env[65107]: _type = "Task" [ 883.114379] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.126719] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102936, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.147821] env[65107]: DEBUG nova.network.neutron [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updated VIF entry in instance network info cache for port 8bb68bf5-be11-4086-a291-e8d4eeb2047b. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 883.148843] env[65107]: DEBUG nova.network.neutron [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updating instance_info_cache with network_info: [{"id": "8bb68bf5-be11-4086-a291-e8d4eeb2047b", "address": "fa:16:3e:6b:29:71", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": null, "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8bb68bf5-be", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 883.176567] env[65107]: WARNING neutronclient.v2_0.client [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.177296] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.177682] env[65107]: WARNING openstack [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.492404] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563405} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.492726] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e/64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.492960] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.493238] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd6b76b0-41cd-4d99-873c-03fdae09d984 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.500641] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 883.500641] env[65107]: value = "task-5102937" [ 883.500641] env[65107]: _type = "Task" [ 883.500641] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.512222] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102937, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.585435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.695s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.585996] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 883.592451] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.832s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.592451] env[65107]: INFO nova.compute.claims [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.626814] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102936, 'name': Rename_Task, 'duration_secs': 0.445495} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.627119] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.627392] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9294b657-f4a1-407f-8a72-0648c398ae78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.636431] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 883.636431] env[65107]: value = "task-5102938" [ 883.636431] env[65107]: _type = "Task" [ 883.636431] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.644574] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.652758] env[65107]: DEBUG oslo_concurrency.lockutils [req-75353521-b89e-4fee-b709-6a731b005ea9 req-c7314eaf-e19f-498c-b6e3-dc7ab00237e0 service nova] Releasing lock "refresh_cache-701a1a16-9348-4748-bb24-ee245a76566c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.788969] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "4a42392a-a4b4-43b7-a410-0b69441435be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.789255] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "4a42392a-a4b4-43b7-a410-0b69441435be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 883.948453] env[65107]: DEBUG nova.network.neutron [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Updating instance_info_cache with network_info: [{"id": "875128de-2b86-4f3c-bef3-92733849587c", "address": "fa:16:3e:55:fd:2a", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap875128de-2b", "ovs_interfaceid": "875128de-2b86-4f3c-bef3-92733849587c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 883.968892] env[65107]: WARNING openstack [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.968892] env[65107]: WARNING openstack [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.014750] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102937, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069653} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.015167] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.016113] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04589da2-5f08-456f-a4df-4f4d716e7bce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.045629] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e/64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.045882] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb57676a-41ec-4ead-8c05-182be314ef2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.069301] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 884.069301] env[65107]: value = "task-5102939" [ 884.069301] env[65107]: _type = "Task" [ 884.069301] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.080297] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102939, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.099213] env[65107]: DEBUG nova.compute.utils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 884.101841] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 884.102188] env[65107]: DEBUG nova.network.neutron [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 884.102682] env[65107]: WARNING neutronclient.v2_0.client [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.103053] env[65107]: WARNING neutronclient.v2_0.client [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.103816] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.104217] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.147771] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102938, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.292883] env[65107]: DEBUG nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 884.454018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.454018] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Instance network_info: |[{"id": "875128de-2b86-4f3c-bef3-92733849587c", "address": "fa:16:3e:55:fd:2a", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap875128de-2b", "ovs_interfaceid": "875128de-2b86-4f3c-bef3-92733849587c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 884.454018] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:fd:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '875128de-2b86-4f3c-bef3-92733849587c', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.463138] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 884.463525] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.464786] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc6b7763-af26-4e60-a9bd-32bc2ec54772 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.488637] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.488637] env[65107]: value = "task-5102940" [ 884.488637] env[65107]: _type = "Task" [ 884.488637] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.501009] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102940, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.581506] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102939, 'name': ReconfigVM_Task, 'duration_secs': 0.361478} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.581506] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e/64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.582276] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9380c97d-b2c9-4e8e-a6b3-9fc92f9e7ba8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.590130] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 884.590130] env[65107]: value = "task-5102941" [ 884.590130] env[65107]: _type = "Task" [ 884.590130] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.601165] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 884.606230] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102941, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.648136] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102938, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.826896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.835551] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.836315] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.836669] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.920672] env[65107]: DEBUG nova.policy [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '531aa60d7da4435c8a7e59f02954e20b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7db97cf40794c5196e31ab5fa8a0477', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 884.977399] env[65107]: DEBUG nova.network.neutron [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updating instance_info_cache with network_info: [{"id": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "address": "fa:16:3e:39:cf:6d", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap190c1749-3f", "ovs_interfaceid": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 884.988207] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received event network-vif-plugged-c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 884.988207] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Acquiring lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.988487] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.988625] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.988882] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] No waiting events found dispatching network-vif-plugged-c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 884.989078] env[65107]: WARNING nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received unexpected event network-vif-plugged-c2fdc887-a837-4102-9f3a-85fec34ad2b9 for instance with vm_state building and task_state spawning. [ 884.989275] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received event network-changed-c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 884.990383] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Refreshing instance network info cache due to event network-changed-c2fdc887-a837-4102-9f3a-85fec34ad2b9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 884.990628] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Acquiring lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.006694] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102940, 'name': CreateVM_Task, 'duration_secs': 0.420491} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.006926] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.007611] env[65107]: WARNING neutronclient.v2_0.client [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.007988] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.008171] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.008574] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 885.009179] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fc22715-a86d-4518-95cd-0ce3d8ecef69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.019557] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 885.019557] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5236d0bb-7a7f-827f-462d-480bbfc5ee45" [ 885.019557] env[65107]: _type = "Task" [ 885.019557] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.033646] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5236d0bb-7a7f-827f-462d-480bbfc5ee45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.078544] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b783a37-a169-4f4b-9250-9788fe9ec417 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.087884] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adfd903-20e1-4bc2-bb2f-cd5a3a5fdda0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.100933] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102941, 'name': Rename_Task, 'duration_secs': 0.178152} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.132070] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.137459] env[65107]: DEBUG nova.network.neutron [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Successfully updated port: c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 885.138797] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b0c4afc-ef30-4e40-9a59-35557cb8da59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.141308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc0427b-07c6-4173-83c6-de5701dc31cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.156380] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102938, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.161786] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 885.161786] env[65107]: value = "task-5102942" [ 885.161786] env[65107]: _type = "Task" [ 885.161786] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.164681] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79803c3d-5b14-4d18-9bee-551fa2f3082c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.179248] env[65107]: WARNING neutronclient.v2_0.client [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.180266] env[65107]: WARNING openstack [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.180811] env[65107]: WARNING openstack [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.217068] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.217589] env[65107]: DEBUG nova.compute.provider_tree [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.370685] env[65107]: DEBUG nova.network.neutron [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.482650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.483138] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Instance network_info: |[{"id": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "address": "fa:16:3e:39:cf:6d", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap190c1749-3f", "ovs_interfaceid": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 885.483669] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:cf:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4020f51-6e46-4b73-a79e-9fe3fd51b917', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '190c1749-3fb8-4199-b8e7-1ab227d2b30f', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.495146] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 885.495424] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.495692] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4955b99a-223e-4942-b5bd-66d0c6218b80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.522034] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.522034] env[65107]: value = "task-5102943" [ 885.522034] env[65107]: _type = "Task" [ 885.522034] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.546724] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5236d0bb-7a7f-827f-462d-480bbfc5ee45, 'name': SearchDatastore_Task, 'duration_secs': 0.016356} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.549800] env[65107]: DEBUG nova.network.neutron [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Updated VIF entry in instance network info cache for port 33f13ff6-7d02-4931-bda5-04c0658cfd6c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 885.550483] env[65107]: DEBUG nova.network.neutron [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Updating instance_info_cache with network_info: [{"id": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "address": "fa:16:3e:85:cd:c3", "network": {"id": "cf8b5cec-0dbb-48c0-8344-638b7845cd5b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1028060705", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f13ff6-7d", "ovs_interfaceid": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "address": "fa:16:3e:66:c6:76", "network": {"id": "31093185-4798-44f4-a16c-a1b6b13e74ce", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-366736123", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fdc887-a8", "ovs_interfaceid": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.557808] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.558793] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.558793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.558793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.558793] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.560328] env[65107]: DEBUG nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Received event network-changed-4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 885.560572] env[65107]: DEBUG nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Refreshing instance network info cache due to event network-changed-4659264c-9645-4cc9-a6aa-4d3299ae4c31. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 885.560715] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Acquiring lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.560855] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Acquired lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.561024] env[65107]: DEBUG nova.network.neutron [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Refreshing network info cache for port 4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 885.562553] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102943, 'name': CreateVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.563146] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-709c8db9-71cc-47cd-95ea-2c2c04cac0b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.575375] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.575569] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.576343] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9aa4dc25-2c21-454f-8ebd-ae9f3164c149 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.584552] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 885.584552] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fccddc-7078-4fdf-195b-517de26a6f3d" [ 885.584552] env[65107]: _type = "Task" [ 885.584552] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.599090] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fccddc-7078-4fdf-195b-517de26a6f3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.640943] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 885.645973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "refresh_cache-fb4229c6-b80d-49e9-a68b-baa12c0004ed" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.646271] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquired lock "refresh_cache-fb4229c6-b80d-49e9-a68b-baa12c0004ed" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.646548] env[65107]: DEBUG nova.network.neutron [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 885.660747] env[65107]: DEBUG oslo_vmware.api [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102938, 'name': PowerOnVM_Task, 'duration_secs': 1.554783} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.661723] env[65107]: DEBUG nova.network.neutron [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Successfully created port: b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 885.665942] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.666222] env[65107]: INFO nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Took 12.93 seconds to spawn the instance on the hypervisor. [ 885.666473] env[65107]: DEBUG nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 885.669704] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07cda82-8ffe-4407-bb1b-36bd2bf77ace {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.690575] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 885.690575] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 885.690575] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 885.691188] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 885.691188] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 885.691188] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 885.691188] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.691552] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 885.691552] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 885.691696] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 885.691945] env[65107]: DEBUG nova.virt.hardware [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 885.692312] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102942, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.693531] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b910b4-415b-4827-80c1-7d4c24a69d2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.705167] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3271a803-0e0c-4758-b57b-0b4a2b8cdee0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.725679] env[65107]: DEBUG nova.scheduler.client.report [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 885.873973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.045044] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102943, 'name': CreateVM_Task, 'duration_secs': 0.4898} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.045888] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 886.052616] env[65107]: WARNING neutronclient.v2_0.client [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.052616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.052616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.052616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 886.052616] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca001ee8-e750-4e9a-b686-f9e515d5be9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.057732] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 886.057732] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522ecf8c-0b59-0f60-8147-9bf4d19ce892" [ 886.057732] env[65107]: _type = "Task" [ 886.057732] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.067896] env[65107]: DEBUG oslo_concurrency.lockutils [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Releasing lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.068231] env[65107]: DEBUG nova.compute.manager [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Received event network-vif-plugged-4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 886.068543] env[65107]: DEBUG oslo_concurrency.lockutils [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Acquiring lock "40e17da0-b459-4619-8e9e-18c3093328c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 886.068839] env[65107]: DEBUG oslo_concurrency.lockutils [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Lock "40e17da0-b459-4619-8e9e-18c3093328c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 886.069148] env[65107]: DEBUG oslo_concurrency.lockutils [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] Lock "40e17da0-b459-4619-8e9e-18c3093328c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.069523] env[65107]: DEBUG nova.compute.manager [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] No waiting events found dispatching network-vif-plugged-4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 886.071163] env[65107]: WARNING nova.compute.manager [req-317edc63-bac6-4619-9093-43bf17674bf5 req-4c967bbc-7873-4a09-a338-f5ed367ed89e service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Received unexpected event network-vif-plugged-4659264c-9645-4cc9-a6aa-4d3299ae4c31 for instance with vm_state building and task_state spawning. [ 886.071163] env[65107]: WARNING neutronclient.v2_0.client [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.071311] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.071784] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.081789] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522ecf8c-0b59-0f60-8147-9bf4d19ce892, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.081789] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Acquired lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.081789] env[65107]: DEBUG nova.network.neutron [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Refreshing network info cache for port c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 886.101013] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fccddc-7078-4fdf-195b-517de26a6f3d, 'name': SearchDatastore_Task, 'duration_secs': 0.011143} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.103208] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e726e73c-23b4-4a42-8a99-b0bde0fa4fc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.112192] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 886.112192] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52284667-4578-b903-19ae-4bf22252e0a7" [ 886.112192] env[65107]: _type = "Task" [ 886.112192] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.125667] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52284667-4578-b903-19ae-4bf22252e0a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.154848] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.155588] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.177665] env[65107]: DEBUG oslo_vmware.api [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102942, 'name': PowerOnVM_Task, 'duration_secs': 0.612717} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.177881] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.178120] env[65107]: INFO nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Took 16.29 seconds to spawn the instance on the hypervisor. [ 886.178305] env[65107]: DEBUG nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 886.179521] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869177f5-2bc0-4ccb-b9ad-b92ff17a15a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.204198] env[65107]: INFO nova.compute.manager [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Took 57.03 seconds to build instance. [ 886.231960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.232585] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 886.236192] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.927s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 886.236413] env[65107]: DEBUG nova.objects.instance [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lazy-loading 'resources' on Instance uuid 2d0a93a4-59a8-4788-93ce-c34af1432a4f {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.444648] env[65107]: DEBUG nova.network.neutron [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 886.569289] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522ecf8c-0b59-0f60-8147-9bf4d19ce892, 'name': SearchDatastore_Task, 'duration_secs': 0.013347} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.569640] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.569892] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.570127] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.584881] env[65107]: WARNING neutronclient.v2_0.client [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.585032] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.585388] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.626282] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52284667-4578-b903-19ae-4bf22252e0a7, 'name': SearchDatastore_Task, 'duration_secs': 0.01222} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.626575] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.626844] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 3f34d6a2-0370-4239-9556-b4333e454d58/3f34d6a2-0370-4239-9556-b4333e454d58.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.627155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.627343] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.627568] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-003a1f99-467f-4147-ab91-2944b0a15daa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.629745] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1380369-a47d-4980-a936-dd68f2523038 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.639584] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 886.639584] env[65107]: value = "task-5102944" [ 886.639584] env[65107]: _type = "Task" [ 886.639584] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.644611] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.644816] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.646268] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41fb51e8-c084-43fe-abb6-9304705651f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.652354] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.656368] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 886.656368] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c12acb-f207-db6d-9ec7-4e7dc45340fd" [ 886.656368] env[65107]: _type = "Task" [ 886.656368] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.665710] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c12acb-f207-db6d-9ec7-4e7dc45340fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.697815] env[65107]: INFO nova.compute.manager [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Took 60.40 seconds to build instance. [ 886.706884] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86e80378-768c-42c3-aebf-13670865a07a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "40e17da0-b459-4619-8e9e-18c3093328c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.987s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.740305] env[65107]: DEBUG nova.compute.utils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 886.745699] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 886.746182] env[65107]: DEBUG nova.network.neutron [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 886.746293] env[65107]: WARNING neutronclient.v2_0.client [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.746774] env[65107]: WARNING neutronclient.v2_0.client [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.748175] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.748175] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.147223] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.147727] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.162477] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102944, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.173198] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c12acb-f207-db6d-9ec7-4e7dc45340fd, 'name': SearchDatastore_Task, 'duration_secs': 0.036302} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.175478] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6002ed1-0663-4e82-8990-f7f9c0269ee0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.178200] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84fb0ab3-a3ea-4a5b-a96b-055466e60d0c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.187987] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6081c7f1-be7f-4212-b68d-ab06d05e3366 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.193145] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 887.193145] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52438ea6-536b-6d48-b86d-bf6787de3388" [ 887.193145] env[65107]: _type = "Task" [ 887.193145] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.227839] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0cddf89a-47a9-4351-a513-1513a15b6ef9 tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.938s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.228895] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527d40b1-fcda-4101-8ad1-e414e9d98dbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.236389] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52438ea6-536b-6d48-b86d-bf6787de3388, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.240730] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be64ddb-02b5-43cc-8fe2-9f21dcae817a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.247212] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 887.259246] env[65107]: DEBUG nova.compute.provider_tree [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.306490] env[65107]: DEBUG nova.policy [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32155078a41249a297470fd4083860df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f9188f3e35e4d06a977e2180918c616', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 887.385354] env[65107]: DEBUG nova.network.neutron [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Successfully updated port: b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 887.391268] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f919321-38dc-4727-a600-81a702272ed9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.413281] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 887.418886] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.418886] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.652909] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.799817} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.653220] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 3f34d6a2-0370-4239-9556-b4333e454d58/3f34d6a2-0370-4239-9556-b4333e454d58.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.654091] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.654091] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cfe4d6a-ea54-4727-ba03-23adb61f940b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.662500] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 887.662500] env[65107]: value = "task-5102945" [ 887.662500] env[65107]: _type = "Task" [ 887.662500] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.673057] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102945, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.706269] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52438ea6-536b-6d48-b86d-bf6787de3388, 'name': SearchDatastore_Task, 'duration_secs': 0.072469} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.706269] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.706269] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 9fa949ea-358c-46e7-a0f2-4c3275493b64/9fa949ea-358c-46e7-a0f2-4c3275493b64.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.706269] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-446bd18a-0187-4e00-a287-74c67a3784cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.714153] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 887.714153] env[65107]: value = "task-5102946" [ 887.714153] env[65107]: _type = "Task" [ 887.714153] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.723304] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102946, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.766161] env[65107]: DEBUG nova.scheduler.client.report [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 887.830805] env[65107]: DEBUG nova.network.neutron [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Successfully created port: 355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 887.864934] env[65107]: WARNING neutronclient.v2_0.client [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.865654] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.866046] env[65107]: WARNING openstack [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.891545] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.891908] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquired lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.892059] env[65107]: DEBUG nova.network.neutron [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 887.928750] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.929238] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5022709c-bd6d-417f-b372-b22fe625f14f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.944186] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 887.944186] env[65107]: value = "task-5102947" [ 887.944186] env[65107]: _type = "Task" [ 887.944186] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.956880] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 887.957243] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 887.963371] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.963898] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.023708] env[65107]: WARNING neutronclient.v2_0.client [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.023968] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.024890] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.175467] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102945, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115744} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.175825] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.176668] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a57783-7e1b-4d5a-856a-9b42fb8da9e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.205390] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 3f34d6a2-0370-4239-9556-b4333e454d58/3f34d6a2-0370-4239-9556-b4333e454d58.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.205390] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9f95b13-c8ae-4801-b4f6-71e9622d90cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.230294] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102946, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47823} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.231905] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 9fa949ea-358c-46e7-a0f2-4c3275493b64/9fa949ea-358c-46e7-a0f2-4c3275493b64.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 888.232148] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.232650] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 888.232650] env[65107]: value = "task-5102948" [ 888.232650] env[65107]: _type = "Task" [ 888.232650] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.232835] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-89954ef1-f40c-4949-8b94-944e6d995983 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.248852] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.250631] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 888.250631] env[65107]: value = "task-5102949" [ 888.250631] env[65107]: _type = "Task" [ 888.250631] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.261722] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.271938] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 888.275938] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.278359] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.934s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.310697] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 888.310945] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 888.311117] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 888.312589] env[65107]: DEBUG nova.virt.hardware [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 888.313808] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485aebd8-9af2-4f69-b109-c73bfc2c9dd3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.319802] env[65107]: INFO nova.scheduler.client.report [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted allocations for instance 2d0a93a4-59a8-4788-93ce-c34af1432a4f [ 888.330479] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa38c60f-105b-4729-9e70-35c5e3c0c13b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.396585] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.396992] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.476779] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 888.477676] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 888.477676] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 888.477676] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 888.477676] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 888.477676] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 888.477931] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.477984] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 888.478636] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 888.478636] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 888.478636] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 888.486649] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d668eb1f-d88a-4136-99ae-4cd4dd691c76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.505355] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 888.505355] env[65107]: value = "task-5102950" [ 888.505355] env[65107]: _type = "Task" [ 888.505355] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.517651] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102950, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.693124] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc89dd0b-6d7d-468c-a131-37bbe7b0ae4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.705817] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09b42a1-6a9a-4f51-ba2d-8862245c581c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.746181] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a79f6cc-d5be-4f96-8047-088f96e88ca6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.756744] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102948, 'name': ReconfigVM_Task, 'duration_secs': 0.519211} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.759265] env[65107]: DEBUG nova.network.neutron [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Updating instance_info_cache with network_info: [{"id": "c23ac068-c418-45e9-8edc-0e541664cbef", "address": "fa:16:3e:ac:ff:7c", "network": {"id": "9df14d45-79e5-4ae3-954b-8d559953a7c0", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-674583965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79f6cf60cda141238e1b8ae1a1a32c22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6a6f7bb-6106-4cfd-9aef-b85628d0cefa", "external-id": "nsx-vlan-transportzone-194", "segmentation_id": 194, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc23ac068-c4", "ovs_interfaceid": "c23ac068-c418-45e9-8edc-0e541664cbef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 888.763612] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 3f34d6a2-0370-4239-9556-b4333e454d58/3f34d6a2-0370-4239-9556-b4333e454d58.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.765289] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d47969-13ba-4f24-9329-87d9b220f33e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.769548] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d1833aa-9b97-4012-b3a8-f84312d5a4fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.784900] env[65107]: DEBUG nova.compute.provider_tree [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.790616] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.201175} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.790895] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 888.790895] env[65107]: value = "task-5102951" [ 888.790895] env[65107]: _type = "Task" [ 888.790895] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.791454] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.792265] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2749c859-a4fc-4eb6-b757-f2ccdf8b2929 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.804915] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102951, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.824301] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 9fa949ea-358c-46e7-a0f2-4c3275493b64/9fa949ea-358c-46e7-a0f2-4c3275493b64.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.825057] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54c518fb-10f7-4d0e-b76b-3fed2af62dc6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.846397] env[65107]: WARNING neutronclient.v2_0.client [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.847149] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.847855] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.856498] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9f207b2b-cc9f-4b73-a3d8-e413b1f42f5c tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "2d0a93a4-59a8-4788-93ce-c34af1432a4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.914s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.868439] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 888.868439] env[65107]: value = "task-5102952" [ 888.868439] env[65107]: _type = "Task" [ 888.868439] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.883286] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102952, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.016371] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102950, 'name': ReconfigVM_Task, 'duration_secs': 0.187207} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.017166] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 889.180683] env[65107]: DEBUG nova.network.neutron [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Updated VIF entry in instance network info cache for port 4659264c-9645-4cc9-a6aa-4d3299ae4c31. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 889.181159] env[65107]: DEBUG nova.network.neutron [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Updating instance_info_cache with network_info: [{"id": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "address": "fa:16:3e:d4:a3:9f", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4659264c-96", "ovs_interfaceid": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.211112] env[65107]: DEBUG nova.network.neutron [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 889.270993] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Releasing lock "refresh_cache-fb4229c6-b80d-49e9-a68b-baa12c0004ed" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.271525] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Instance network_info: |[{"id": "c23ac068-c418-45e9-8edc-0e541664cbef", "address": "fa:16:3e:ac:ff:7c", "network": {"id": "9df14d45-79e5-4ae3-954b-8d559953a7c0", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-674583965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79f6cf60cda141238e1b8ae1a1a32c22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6a6f7bb-6106-4cfd-9aef-b85628d0cefa", "external-id": "nsx-vlan-transportzone-194", "segmentation_id": 194, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc23ac068-c4", "ovs_interfaceid": "c23ac068-c418-45e9-8edc-0e541664cbef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 889.272169] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:ff:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6a6f7bb-6106-4cfd-9aef-b85628d0cefa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c23ac068-c418-45e9-8edc-0e541664cbef', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.280872] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Creating folder: Project (79f6cf60cda141238e1b8ae1a1a32c22). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.281202] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35ffcd5e-4acd-4fb5-bb87-d9b01c77459b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.292638] env[65107]: DEBUG nova.scheduler.client.report [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.295668] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Created folder: Project (79f6cf60cda141238e1b8ae1a1a32c22) in parent group-v992574. [ 889.295993] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Creating folder: Instances. Parent ref: group-v992764. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.296389] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c2ccd43-8108-45be-be1f-731439058330 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.308027] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102951, 'name': Rename_Task, 'duration_secs': 0.349406} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.309236] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.309236] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5c5502c-a268-430e-a524-dc1e545a333a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.311816] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Created folder: Instances in parent group-v992764. [ 889.312093] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 889.312700] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.312978] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32345e20-e556-448f-8985-83260facc0a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.330355] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 889.330355] env[65107]: value = "task-5102955" [ 889.330355] env[65107]: _type = "Task" [ 889.330355] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.336253] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.336253] env[65107]: value = "task-5102956" [ 889.336253] env[65107]: _type = "Task" [ 889.336253] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.344435] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102955, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.351093] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102956, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.382275] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102952, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.528536] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 889.528807] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 889.528807] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 889.529070] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 889.529271] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 889.529470] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 889.529811] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.530057] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 889.530307] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 889.530635] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 889.530757] env[65107]: DEBUG nova.virt.hardware [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 889.537277] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Reconfiguring VM instance instance-0000001c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 889.537720] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0651a663-2041-415b-ad4d-8145334e8f49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.559520] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "3a975793-f480-4f54-85c9-2c2aea11c18a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.559828] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.565295] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 889.565295] env[65107]: value = "task-5102957" [ 889.565295] env[65107]: _type = "Task" [ 889.565295] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.576319] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.683931] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Releasing lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.683931] env[65107]: DEBUG nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Received event network-vif-plugged-190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 889.684201] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Acquiring lock "9fa949ea-358c-46e7-a0f2-4c3275493b64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.684421] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.684589] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.684760] env[65107]: DEBUG nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] No waiting events found dispatching network-vif-plugged-190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 889.684932] env[65107]: WARNING nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Received unexpected event network-vif-plugged-190c1749-3fb8-4199-b8e7-1ab227d2b30f for instance with vm_state building and task_state spawning. [ 889.685117] env[65107]: DEBUG nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Received event network-changed-190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 889.685278] env[65107]: DEBUG nova.compute.manager [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Refreshing instance network info cache due to event network-changed-190c1749-3fb8-4199-b8e7-1ab227d2b30f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 889.685465] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Acquiring lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.685603] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Acquired lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.685786] env[65107]: DEBUG nova.network.neutron [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Refreshing network info cache for port 190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 889.723713] env[65107]: DEBUG nova.network.neutron [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Successfully updated port: 355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 889.743111] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.743603] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.797718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.519s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.798033] env[65107]: INFO nova.compute.manager [None req-1066006a-eaf3-4c52-ad3c-04e6ab05280b tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Successfully reverted task state from rebuilding on failure for instance. [ 889.806411] env[65107]: DEBUG nova.network.neutron [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Updated VIF entry in instance network info cache for port c2fdc887-a837-4102-9f3a-85fec34ad2b9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 889.806827] env[65107]: DEBUG nova.network.neutron [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Updating instance_info_cache with network_info: [{"id": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "address": "fa:16:3e:85:cd:c3", "network": {"id": "cf8b5cec-0dbb-48c0-8344-638b7845cd5b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1028060705", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.128", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b107fab-ee71-47db-ad4d-3c6f05546843", "external-id": "cl2-zone-554", "segmentation_id": 554, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f13ff6-7d", "ovs_interfaceid": "33f13ff6-7d02-4931-bda5-04c0658cfd6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "address": "fa:16:3e:66:c6:76", "network": {"id": "31093185-4798-44f4-a16c-a1b6b13e74ce", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-366736123", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d1ef40d5301a473ea0a1691fecf01fd2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "669e4919-e0ad-4e23-9f23-4c5f2be0d858", "external-id": "nsx-vlan-transportzone-362", "segmentation_id": 362, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2fdc887-a8", "ovs_interfaceid": "c2fdc887-a837-4102-9f3a-85fec34ad2b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.808350] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.346s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.808569] env[65107]: DEBUG nova.objects.instance [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lazy-loading 'resources' on Instance uuid a4f43fbc-63e3-4b69-bdb1-7aaf47aae090 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.848256] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102955, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.852816] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102956, 'name': CreateVM_Task, 'duration_secs': 0.370142} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.853092] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 889.853655] env[65107]: WARNING neutronclient.v2_0.client [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.854075] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.854257] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.854586] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 889.854851] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47b62463-4e1c-4704-aacd-86447dabad3f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.863098] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 889.863098] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e4ea85-88ff-a6dc-a3df-7608c22f6bc1" [ 889.863098] env[65107]: _type = "Task" [ 889.863098] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.874045] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e4ea85-88ff-a6dc-a3df-7608c22f6bc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.888443] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102952, 'name': ReconfigVM_Task, 'duration_secs': 0.682702} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.888889] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 9fa949ea-358c-46e7-a0f2-4c3275493b64/9fa949ea-358c-46e7-a0f2-4c3275493b64.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.889421] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae1a7380-68dc-4d96-a25a-5c15ef7e2f80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.898966] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 889.898966] env[65107]: value = "task-5102958" [ 889.898966] env[65107]: _type = "Task" [ 889.898966] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.910224] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102958, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.046582] env[65107]: WARNING neutronclient.v2_0.client [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.046861] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.048215] env[65107]: WARNING openstack [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.063863] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 890.079584] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102957, 'name': ReconfigVM_Task, 'duration_secs': 0.349632} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.080885] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Reconfigured VM instance instance-0000001c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 890.082127] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f42042c-a606-467d-a65e-a2c1b821c546 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.113766] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96/71e6e280-50c3-4655-a6a3-40ea4c650d96.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.113766] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ba80b94-d00d-4db4-950c-3fac24c0e23d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.134755] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 890.134755] env[65107]: value = "task-5102959" [ 890.134755] env[65107]: _type = "Task" [ 890.134755] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.145721] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102959, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.188629] env[65107]: WARNING neutronclient.v2_0.client [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.189384] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.189766] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.225960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "refresh_cache-ac631507-8abf-4019-bdd5-7684dd51d1e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.226238] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "refresh_cache-ac631507-8abf-4019-bdd5-7684dd51d1e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.226420] env[65107]: DEBUG nova.network.neutron [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 890.319158] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Releasing lock "refresh_cache-64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.319461] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Received event network-vif-plugged-875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 890.320072] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Acquiring lock "3f34d6a2-0370-4239-9556-b4333e454d58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.320072] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Lock "3f34d6a2-0370-4239-9556-b4333e454d58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.320072] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Lock "3f34d6a2-0370-4239-9556-b4333e454d58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.320294] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] No waiting events found dispatching network-vif-plugged-875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 890.320534] env[65107]: WARNING nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Received unexpected event network-vif-plugged-875128de-2b86-4f3c-bef3-92733849587c for instance with vm_state building and task_state spawning. [ 890.320738] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Received event network-changed-875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 890.320944] env[65107]: DEBUG nova.compute.manager [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Refreshing instance network info cache due to event network-changed-875128de-2b86-4f3c-bef3-92733849587c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 890.324849] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Acquiring lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.325188] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Acquired lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.325482] env[65107]: DEBUG nova.network.neutron [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Refreshing network info cache for port 875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 890.342429] env[65107]: DEBUG oslo_vmware.api [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5102955, 'name': PowerOnVM_Task, 'duration_secs': 0.73833} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.343394] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.343609] env[65107]: INFO nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Took 14.08 seconds to spawn the instance on the hypervisor. [ 890.343790] env[65107]: DEBUG nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 890.344625] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7f0ac5-d114-4480-86f1-052b96ec2fd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.383790] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e4ea85-88ff-a6dc-a3df-7608c22f6bc1, 'name': SearchDatastore_Task, 'duration_secs': 0.011267} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.384331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.384429] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.384641] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.384814] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.384955] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.385254] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0adebfe6-313b-4bde-9748-24f3f09af4b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.399165] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.399481] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.400589] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56a8ce45-bf51-43c3-8bf7-9659e1c6d424 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.405838] env[65107]: DEBUG nova.network.neutron [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updating instance_info_cache with network_info: [{"id": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "address": "fa:16:3e:18:a3:54", "network": {"id": "39ea7370-9401-4188-a7d0-a9b311fc29aa", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-53213752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "f7db97cf40794c5196e31ab5fa8a0477", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3279b33-90", "ovs_interfaceid": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 890.421421] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102958, 'name': Rename_Task, 'duration_secs': 0.193712} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.421608] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 890.421608] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52abbc44-138d-019f-9f97-d519d913dfd5" [ 890.421608] env[65107]: _type = "Task" [ 890.421608] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.421816] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.422330] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a023a26-7bcd-4dc7-8cdc-51aefe89779d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.436937] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52abbc44-138d-019f-9f97-d519d913dfd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.439806] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 890.439806] env[65107]: value = "task-5102960" [ 890.439806] env[65107]: _type = "Task" [ 890.439806] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.450270] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.563981] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "9a46ff51-4035-4709-ae6c-0fb5397200ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.564164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.564394] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "9a46ff51-4035-4709-ae6c-0fb5397200ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.564563] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.564811] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.569079] env[65107]: INFO nova.compute.manager [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Terminating instance [ 890.590115] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.646067] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102959, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.730880] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.730880] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.785672] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e6426c-bd37-458b-ab5b-6849adbcda16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.795510] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7434555b-c4d3-4b9f-92b9-1cc25dfabfea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.830488] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.830895] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.838686] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d33ada-a8f2-430f-9d04-b7a5c773a233 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.841894] env[65107]: WARNING neutronclient.v2_0.client [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.842607] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.843556] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.864828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd866e69-8d99-4104-98ae-745daf1b7e4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.871350] env[65107]: INFO nova.compute.manager [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Took 50.35 seconds to build instance. [ 890.882321] env[65107]: DEBUG nova.compute.provider_tree [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.898919] env[65107]: DEBUG nova.network.neutron [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 890.913283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Releasing lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.913283] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Instance network_info: |[{"id": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "address": "fa:16:3e:18:a3:54", "network": {"id": "39ea7370-9401-4188-a7d0-a9b311fc29aa", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-53213752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "f7db97cf40794c5196e31ab5fa8a0477", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3279b33-90", "ovs_interfaceid": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 890.913669] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:a3:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3279b33-9092-457f-88cc-4c2cc95e6cdc', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.921449] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Creating folder: Project (f7db97cf40794c5196e31ab5fa8a0477). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.922225] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46738929-23d1-4abf-a4b7-d3886ee645ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.934921] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52abbc44-138d-019f-9f97-d519d913dfd5, 'name': SearchDatastore_Task, 'duration_secs': 0.026377} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.937058] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae6910b4-1e8a-4276-9aea-e8cf69370156 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.940164] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Created folder: Project (f7db97cf40794c5196e31ab5fa8a0477) in parent group-v992574. [ 890.940383] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Creating folder: Instances. Parent ref: group-v992767. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.941099] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d91745a4-0a53-4105-bcf4-71de483fe4fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.947990] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 890.947990] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52edd73f-7d9e-9c40-7c01-cea0c5e62b0a" [ 890.947990] env[65107]: _type = "Task" [ 890.947990] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.954296] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102960, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.956822] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Created folder: Instances in parent group-v992767. [ 890.957125] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 890.957334] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.957556] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a6371e6-957e-4051-80d6-252b5c2eea6a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.978331] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52edd73f-7d9e-9c40-7c01-cea0c5e62b0a, 'name': SearchDatastore_Task, 'duration_secs': 0.011073} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.978331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.978331] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] fb4229c6-b80d-49e9-a68b-baa12c0004ed/fb4229c6-b80d-49e9-a68b-baa12c0004ed.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.978331] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56a625d1-8ed4-4655-8f5a-5293830dd578 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.982104] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.982104] env[65107]: value = "task-5102963" [ 890.982104] env[65107]: _type = "Task" [ 890.982104] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.987147] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 890.987147] env[65107]: value = "task-5102964" [ 890.987147] env[65107]: _type = "Task" [ 890.987147] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.993856] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102963, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.000551] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102964, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.079931] env[65107]: DEBUG nova.compute.manager [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 891.080312] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.081387] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4144112f-6876-4dcb-952d-67d45e77eb44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.091801] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.092198] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2399e79-b8ac-48ee-8ce0-3647b54d9f59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.101408] env[65107]: DEBUG oslo_vmware.api [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 891.101408] env[65107]: value = "task-5102965" [ 891.101408] env[65107]: _type = "Task" [ 891.101408] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.111878] env[65107]: DEBUG oslo_vmware.api [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.148641] env[65107]: DEBUG oslo_vmware.api [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5102959, 'name': ReconfigVM_Task, 'duration_secs': 0.641425} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.149184] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96/71e6e280-50c3-4655-a6a3-40ea4c650d96.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.149635] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 891.373617] env[65107]: DEBUG oslo_concurrency.lockutils [None req-acc19233-4939-4a1c-97a5-6c37d8cc9d6a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3f34d6a2-0370-4239-9556-b4333e454d58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.977s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.386517] env[65107]: DEBUG nova.scheduler.client.report [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.449267] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.449700] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.462129] env[65107]: DEBUG oslo_vmware.api [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5102960, 'name': PowerOnVM_Task, 'duration_secs': 0.798946} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.462129] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.462129] env[65107]: INFO nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Took 13.11 seconds to spawn the instance on the hypervisor. [ 891.462129] env[65107]: DEBUG nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 891.462513] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e849c7e-6794-47a5-941e-c42d9f0b27ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.483144] env[65107]: WARNING neutronclient.v2_0.client [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.484023] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.484824] env[65107]: WARNING openstack [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.508211] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102963, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.511997] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102964, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.613425] env[65107]: DEBUG oslo_vmware.api [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102965, 'name': PowerOffVM_Task, 'duration_secs': 0.335136} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.613778] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.614014] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.614495] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d57cea8-51cf-4f1f-a8dd-d0d68a56813d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.657890] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2c5464-35e4-4d9d-a9cd-707411d0cd49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.678612] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5fa2659-bf1a-44a3-be74-835c92147cae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.703173] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 891.708145] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.708378] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.708561] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleting the datastore file [datastore1] 9a46ff51-4035-4709-ae6c-0fb5397200ab {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.709117] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3439f8af-748e-4ec2-8c70-d544ff96fa53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.718203] env[65107]: DEBUG oslo_vmware.api [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 891.718203] env[65107]: value = "task-5102967" [ 891.718203] env[65107]: _type = "Task" [ 891.718203] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.728623] env[65107]: DEBUG oslo_vmware.api [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.820731] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.821180] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.832804] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Received event network-vif-deleted-fcca9afe-edba-479c-93f6-1310627d946f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 891.833023] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Received event network-vif-plugged-c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 891.833743] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Acquiring lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.833743] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.833972] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.834101] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] No waiting events found dispatching network-vif-plugged-c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 891.834330] env[65107]: WARNING nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Received unexpected event network-vif-plugged-c23ac068-c418-45e9-8edc-0e541664cbef for instance with vm_state building and task_state spawning. [ 891.834602] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Received event network-changed-c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 891.835048] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Refreshing instance network info cache due to event network-changed-c23ac068-c418-45e9-8edc-0e541664cbef. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 891.835048] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Acquiring lock "refresh_cache-fb4229c6-b80d-49e9-a68b-baa12c0004ed" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.835283] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Acquired lock "refresh_cache-fb4229c6-b80d-49e9-a68b-baa12c0004ed" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.835461] env[65107]: DEBUG nova.network.neutron [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Refreshing network info cache for port c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 891.883305] env[65107]: WARNING neutronclient.v2_0.client [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.884028] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.884510] env[65107]: WARNING openstack [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.893701] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.085s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.896614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.025s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.898120] env[65107]: INFO nova.compute.claims [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.925595] env[65107]: INFO nova.scheduler.client.report [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Deleted allocations for instance a4f43fbc-63e3-4b69-bdb1-7aaf47aae090 [ 891.982032] env[65107]: DEBUG nova.network.neutron [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updated VIF entry in instance network info cache for port 190c1749-3fb8-4199-b8e7-1ab227d2b30f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 891.982534] env[65107]: DEBUG nova.network.neutron [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updating instance_info_cache with network_info: [{"id": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "address": "fa:16:3e:39:cf:6d", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap190c1749-3f", "ovs_interfaceid": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 891.988831] env[65107]: INFO nova.compute.manager [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Took 51.21 seconds to build instance. [ 892.011700] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102964, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579158} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.011971] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102963, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.013142] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] fb4229c6-b80d-49e9-a68b-baa12c0004ed/fb4229c6-b80d-49e9-a68b-baa12c0004ed.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.013142] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.013352] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-895abe1b-fd3b-41bd-a7d8-d92f960377ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.023320] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 892.023320] env[65107]: value = "task-5102968" [ 892.023320] env[65107]: _type = "Task" [ 892.023320] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.039375] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.190171] env[65107]: WARNING neutronclient.v2_0.client [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.191486] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.192246] env[65107]: WARNING openstack [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.212193] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.212193] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.231113] env[65107]: DEBUG oslo_vmware.api [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5102967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156546} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.231113] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.231215] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.231370] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.231547] env[65107]: INFO nova.compute.manager [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Took 1.15 seconds to destroy the instance on the hypervisor. [ 892.231861] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 892.232209] env[65107]: DEBUG nova.compute.manager [-] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 892.232345] env[65107]: DEBUG nova.network.neutron [-] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 892.232708] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.233543] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.233906] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.340755] env[65107]: WARNING neutronclient.v2_0.client [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.340755] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.340755] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.355519] env[65107]: DEBUG nova.network.neutron [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Updating instance_info_cache with network_info: [{"id": "355ade04-dfe9-489a-81d1-b7bee4c4833e", "address": "fa:16:3e:de:1f:ee", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355ade04-df", "ovs_interfaceid": "355ade04-dfe9-489a-81d1-b7bee4c4833e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 892.426350] env[65107]: DEBUG nova.network.neutron [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Port 3d38d576-1871-49a5-bc10-aef1a37abfa0 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 892.437761] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a8328e5-4896-4bb6-bb40-c7d5dccf961b tempest-ServersWithSpecificFlavorTestJSON-1123838888 tempest-ServersWithSpecificFlavorTestJSON-1123838888-project-member] Lock "a4f43fbc-63e3-4b69-bdb1-7aaf47aae090" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.757s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.489289] env[65107]: DEBUG oslo_concurrency.lockutils [req-ef521815-aca6-47f0-95dd-0df12f5b011f req-5d798f17-bf39-4389-9d25-bf27a4b2d245 service nova] Releasing lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.491630] env[65107]: DEBUG oslo_concurrency.lockutils [None req-214808c0-3192-4f90-bcaf-fc467a618279 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.149s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.505657] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102963, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.511680] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.538251] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080262} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.538251] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.539288] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d38e14e-3d38-4e2f-9512-87439f9e7fb0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.568833] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] fb4229c6-b80d-49e9-a68b-baa12c0004ed/fb4229c6-b80d-49e9-a68b-baa12c0004ed.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.571352] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d458a87c-fa30-48db-8e66-c2dfb28c15c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.591586] env[65107]: DEBUG nova.network.neutron [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Updated VIF entry in instance network info cache for port 875128de-2b86-4f3c-bef3-92733849587c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 892.592069] env[65107]: DEBUG nova.network.neutron [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Updating instance_info_cache with network_info: [{"id": "875128de-2b86-4f3c-bef3-92733849587c", "address": "fa:16:3e:55:fd:2a", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap875128de-2b", "ovs_interfaceid": "875128de-2b86-4f3c-bef3-92733849587c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 892.596500] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 892.596500] env[65107]: value = "task-5102969" [ 892.596500] env[65107]: _type = "Task" [ 892.596500] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.606758] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102969, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.858987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "refresh_cache-ac631507-8abf-4019-bdd5-7684dd51d1e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.859590] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Instance network_info: |[{"id": "355ade04-dfe9-489a-81d1-b7bee4c4833e", "address": "fa:16:3e:de:1f:ee", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355ade04-df", "ovs_interfaceid": "355ade04-dfe9-489a-81d1-b7bee4c4833e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 892.860255] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:1f:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '355ade04-dfe9-489a-81d1-b7bee4c4833e', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.872951] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 892.874380] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.874510] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4e781e9-d2ae-4a0c-b9c6-175700cf91d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.901215] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.901215] env[65107]: value = "task-5102970" [ 892.901215] env[65107]: _type = "Task" [ 892.901215] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.913734] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102970, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.008185] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102963, 'name': CreateVM_Task, 'duration_secs': 1.524451} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.008515] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 893.012789] env[65107]: WARNING neutronclient.v2_0.client [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.013381] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.013630] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 893.013971] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 893.015833] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4736641c-efd3-441e-9aac-e4a427b48172 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.024058] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 893.024058] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee8c78-64f5-84ea-780e-606c7a2139e4" [ 893.024058] env[65107]: _type = "Task" [ 893.024058] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.037142] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee8c78-64f5-84ea-780e-606c7a2139e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.099335] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f330b12-b6b5-4adc-b5e7-f9781dbe014f req-240c7f6f-21ea-42b2-9508-fac5eebc82cb service nova] Releasing lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 893.111436] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102969, 'name': ReconfigVM_Task, 'duration_secs': 0.479749} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.115049] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Reconfigured VM instance instance-00000042 to attach disk [datastore2] fb4229c6-b80d-49e9-a68b-baa12c0004ed/fb4229c6-b80d-49e9-a68b-baa12c0004ed.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.116124] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f1a99d2-5382-4e84-b035-dfde0460930c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.126733] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 893.126733] env[65107]: value = "task-5102971" [ 893.126733] env[65107]: _type = "Task" [ 893.126733] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.141905] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102971, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.194634] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.195185] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.325754] env[65107]: DEBUG nova.network.neutron [-] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.408198] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2bdd54-c5e3-443f-ae18-9074d2b34b23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.420364] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae738010-7cc1-499a-9ff3-113042b553be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.423826] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102970, 'name': CreateVM_Task, 'duration_secs': 0.482351} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.424363] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 893.426211] env[65107]: WARNING neutronclient.v2_0.client [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.426211] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.468710] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1c3932-ec5d-411b-9e8f-fdc4a9a6fa7e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.478109] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 893.479097] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.479097] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.491134] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ddc13b-1379-4481-b577-e6d121b5e6d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.507273] env[65107]: DEBUG nova.compute.provider_tree [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.534651] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee8c78-64f5-84ea-780e-606c7a2139e4, 'name': SearchDatastore_Task, 'duration_secs': 0.012456} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.534953] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 893.535209] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.535442] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.535581] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 893.535755] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.536049] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 893.536356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 893.536581] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1618e8d2-c4d5-44e4-98da-fe4e26639c0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.538612] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fadd7cb3-fa5e-4627-b5c2-2ac414af898d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.545310] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 893.545310] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ad1ba5-cdea-1adb-e4ec-1c6ef2b28fad" [ 893.545310] env[65107]: _type = "Task" [ 893.545310] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.549487] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.549671] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.550785] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99d658e3-46f2-409c-8532-19d0b5308f7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.556538] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ad1ba5-cdea-1adb-e4ec-1c6ef2b28fad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.560801] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 893.560801] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527278f2-9197-667e-da87-fe00fbd76cf0" [ 893.560801] env[65107]: _type = "Task" [ 893.560801] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.569423] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527278f2-9197-667e-da87-fe00fbd76cf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.641720] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102971, 'name': Rename_Task, 'duration_secs': 0.204534} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.642392] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.642845] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1de32dcd-265c-4f2a-b30e-01dc822f52ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.651780] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 893.651780] env[65107]: value = "task-5102972" [ 893.651780] env[65107]: _type = "Task" [ 893.651780] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.663926] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.832099] env[65107]: INFO nova.compute.manager [-] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Took 1.60 seconds to deallocate network for instance. [ 893.918333] env[65107]: WARNING neutronclient.v2_0.client [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.918333] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.918333] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.013199] env[65107]: DEBUG nova.scheduler.client.report [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 894.062446] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ad1ba5-cdea-1adb-e4ec-1c6ef2b28fad, 'name': SearchDatastore_Task, 'duration_secs': 0.011463} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.066828] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.067114] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.067610] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.080227] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527278f2-9197-667e-da87-fe00fbd76cf0, 'name': SearchDatastore_Task, 'duration_secs': 0.00968} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.080227] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c784a04-4be1-4ca1-bee6-f4db31971067 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.091084] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 894.091084] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5243c0ed-6afb-cbf2-940d-8a96d8b584e9" [ 894.091084] env[65107]: _type = "Task" [ 894.091084] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.102703] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5243c0ed-6afb-cbf2-940d-8a96d8b584e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.128155] env[65107]: INFO nova.compute.manager [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Rescuing [ 894.128155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.128155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.128155] env[65107]: DEBUG nova.network.neutron [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 894.165783] env[65107]: DEBUG oslo_vmware.api [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102972, 'name': PowerOnVM_Task, 'duration_secs': 0.512155} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.166084] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.166369] env[65107]: INFO nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Took 12.74 seconds to spawn the instance on the hypervisor. [ 894.166511] env[65107]: DEBUG nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 894.167448] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e00886-719a-4759-a577-a9ece91aed69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.211063] env[65107]: DEBUG nova.network.neutron [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Updated VIF entry in instance network info cache for port c23ac068-c418-45e9-8edc-0e541664cbef. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 894.211063] env[65107]: DEBUG nova.network.neutron [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Updating instance_info_cache with network_info: [{"id": "c23ac068-c418-45e9-8edc-0e541664cbef", "address": "fa:16:3e:ac:ff:7c", "network": {"id": "9df14d45-79e5-4ae3-954b-8d559953a7c0", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-674583965-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79f6cf60cda141238e1b8ae1a1a32c22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6a6f7bb-6106-4cfd-9aef-b85628d0cefa", "external-id": "nsx-vlan-transportzone-194", "segmentation_id": 194, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc23ac068-c4", "ovs_interfaceid": "c23ac068-c418-45e9-8edc-0e541664cbef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 894.340593] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.483606] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.519929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.521032] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 894.524289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.391s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.524621] env[65107]: DEBUG nova.objects.instance [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lazy-loading 'resources' on Instance uuid d65e98bb-f54d-4803-95ca-0fc9b6c184bc {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.615069] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5243c0ed-6afb-cbf2-940d-8a96d8b584e9, 'name': SearchDatastore_Task, 'duration_secs': 0.012488} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.615069] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.615884] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] cddb14db-34c3-4783-b21e-e0b4623a3a9e/cddb14db-34c3-4783-b21e-e0b4623a3a9e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.616251] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.616462] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.616689] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3259f395-22f6-428c-b233-5da673d110cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.619038] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e6a041b-c5be-42f4-8537-542308eef16c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.628278] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 894.628278] env[65107]: value = "task-5102973" [ 894.628278] env[65107]: _type = "Task" [ 894.628278] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.633028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.633028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.633028] env[65107]: DEBUG nova.network.neutron [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 894.633028] env[65107]: WARNING neutronclient.v2_0.client [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.633028] env[65107]: WARNING openstack [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.633463] env[65107]: WARNING openstack [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.642104] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.642814] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.643308] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.655560] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.656026] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.657132] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eef9aee3-2698-44fa-8c33-ba45b9e291f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.670022] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.670022] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 894.670022] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b3b0b0-f257-70cb-b077-1a61cb498f50" [ 894.670022] env[65107]: _type = "Task" [ 894.670022] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.678152] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b3b0b0-f257-70cb-b077-1a61cb498f50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.688754] env[65107]: INFO nova.compute.manager [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Took 38.53 seconds to build instance. [ 894.714875] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Releasing lock "refresh_cache-fb4229c6-b80d-49e9-a68b-baa12c0004ed" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.715372] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Received event network-vif-plugged-b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 894.715500] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Acquiring lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.715873] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.715873] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.716037] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] No waiting events found dispatching network-vif-plugged-b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 894.716235] env[65107]: WARNING nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Received unexpected event network-vif-plugged-b3279b33-9092-457f-88cc-4c2cc95e6cdc for instance with vm_state building and task_state spawning. [ 894.717835] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Received event network-changed-b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 894.717835] env[65107]: DEBUG nova.compute.manager [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Refreshing instance network info cache due to event network-changed-b3279b33-9092-457f-88cc-4c2cc95e6cdc. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 894.717835] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Acquiring lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.717835] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Acquired lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.717835] env[65107]: DEBUG nova.network.neutron [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Refreshing network info cache for port b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 894.773631] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.773742] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.032248] env[65107]: DEBUG nova.compute.utils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 895.037022] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 895.037433] env[65107]: DEBUG nova.network.neutron [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 895.037682] env[65107]: WARNING neutronclient.v2_0.client [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.037963] env[65107]: WARNING neutronclient.v2_0.client [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.038606] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.038980] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.147924] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102973, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.183282] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b3b0b0-f257-70cb-b077-1a61cb498f50, 'name': SearchDatastore_Task, 'duration_secs': 0.017905} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.187726] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc5cbbdf-7362-43d8-974b-e9500e43d8b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.190609] env[65107]: DEBUG oslo_concurrency.lockutils [None req-90952b21-aa72-4357-a8d7-dda43eb92bcb tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.037s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 895.195460] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 895.195460] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8ffe1-4456-8cbc-6080-e085a3beb8ae" [ 895.195460] env[65107]: _type = "Task" [ 895.195460] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.206336] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8ffe1-4456-8cbc-6080-e085a3beb8ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.221177] env[65107]: WARNING neutronclient.v2_0.client [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.222288] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.222361] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.275991] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 895.547872] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 895.580163] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9302c4-8dd3-416c-a753-9f98266e4454 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.589706] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf68223-4209-4a74-b6ca-d40f59dc8490 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.631016] env[65107]: WARNING openstack [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.631016] env[65107]: WARNING openstack [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.635563] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c2acaf-dab9-40f1-ae6f-405257a6ad52 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.647775] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.816222} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.649964] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] cddb14db-34c3-4783-b21e-e0b4623a3a9e/cddb14db-34c3-4783-b21e-e0b4623a3a9e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 895.650215] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.652320] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.652320] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.658128] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f44d6f55-939e-4344-99e5-6a3aca2377b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.663377] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694ac4e4-4e04-47bb-becc-65a943f5337b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.679404] env[65107]: DEBUG nova.compute.provider_tree [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.682203] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 895.682203] env[65107]: value = "task-5102974" [ 895.682203] env[65107]: _type = "Task" [ 895.682203] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.691293] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102974, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.694276] env[65107]: DEBUG nova.policy [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7d2ccd05c364d29a5e1b2283a65e5d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd4b1120670a4df9af1fb8905c4b2b8f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 895.706494] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8ffe1-4456-8cbc-6080-e085a3beb8ae, 'name': SearchDatastore_Task, 'duration_secs': 0.057764} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.706825] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 895.707833] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ac631507-8abf-4019-bdd5-7684dd51d1e9/ac631507-8abf-4019-bdd5-7684dd51d1e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 895.707833] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48585e61-4a47-45f8-b4fd-c00df4936e10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.716374] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 895.716374] env[65107]: value = "task-5102975" [ 895.716374] env[65107]: _type = "Task" [ 895.716374] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.724798] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102975, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.802204] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 896.185197] env[65107]: DEBUG nova.scheduler.client.report [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 896.202838] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102974, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078773} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.202941] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.208040] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d5b197-6ff2-4c31-9f65-9dc932f9ddcc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.235088] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] cddb14db-34c3-4783-b21e-e0b4623a3a9e/cddb14db-34c3-4783-b21e-e0b4623a3a9e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.239032] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3295c8b1-71e1-49f7-9c53-89c2c252d7f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.263487] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102975, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.265211] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 896.265211] env[65107]: value = "task-5102976" [ 896.265211] env[65107]: _type = "Task" [ 896.265211] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.276256] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102976, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.277571] env[65107]: DEBUG nova.network.neutron [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Successfully created port: 03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 896.308155] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.308459] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.443351] env[65107]: WARNING neutronclient.v2_0.client [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.444513] env[65107]: WARNING openstack [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.445118] env[65107]: WARNING openstack [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.472256] env[65107]: WARNING neutronclient.v2_0.client [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.473301] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.474980] env[65107]: WARNING openstack [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.564172] env[65107]: DEBUG nova.compute.manager [req-d4f169a9-decc-45c9-aa90-75bae336754b req-61a6034c-c237-41cd-875d-2f167fb5a1e3 service nova] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Received event network-vif-deleted-0a1d113a-256a-4a98-a5c6-45d4fe610599 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 896.567471] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 896.613379] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 896.613379] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 896.613379] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 896.614834] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 896.616367] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 896.616717] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 896.617073] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.617832] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 896.618015] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 896.619971] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 896.619971] env[65107]: DEBUG nova.virt.hardware [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 896.619971] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fe032e-3937-402c-89da-2724b28bd631 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.632617] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5e2b7f-36c8-4cd7-a4d8-7c0bad5cfc94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.692111] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.168s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.696430] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.187s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.696997] env[65107]: DEBUG nova.objects.instance [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lazy-loading 'resources' on Instance uuid e3510c31-d6be-4e9f-a0a2-a662123861e9 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.741655] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102975, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.948618} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.741854] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] ac631507-8abf-4019-bdd5-7684dd51d1e9/ac631507-8abf-4019-bdd5-7684dd51d1e9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.741997] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.743303] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2398d454-2dcb-4f5f-b48c-883f4f50d01a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.751390] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 896.751390] env[65107]: value = "task-5102977" [ 896.751390] env[65107]: _type = "Task" [ 896.751390] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.764327] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.778188] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.233435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ea49739-44d9-4f33-8139-1d42630e1041 tempest-ServerActionsV293TestJSON-238779998 tempest-ServerActionsV293TestJSON-238779998-project-member] Lock "d65e98bb-f54d-4803-95ca-0fc9b6c184bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.498s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.256366] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.256607] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.275027] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077645} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.275564] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.276531] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193fc9bf-2d5a-485c-9bfb-851030bf68fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.284021] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102976, 'name': ReconfigVM_Task, 'duration_secs': 0.604828} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.286848] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Reconfigured VM instance instance-00000043 to attach disk [datastore1] cddb14db-34c3-4783-b21e-e0b4623a3a9e/cddb14db-34c3-4783-b21e-e0b4623a3a9e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.287974] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-033cfb1e-6b96-4668-a014-1d21a59d085d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.309537] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] ac631507-8abf-4019-bdd5-7684dd51d1e9/ac631507-8abf-4019-bdd5-7684dd51d1e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.313330] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e4c2301-4cf1-433d-8abe-5820a3f24913 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.330400] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 897.330400] env[65107]: value = "task-5102978" [ 897.330400] env[65107]: _type = "Task" [ 897.330400] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.339079] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 897.339079] env[65107]: value = "task-5102979" [ 897.339079] env[65107]: _type = "Task" [ 897.339079] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.349346] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102978, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.355493] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.667780] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e61511-7955-415b-8f27-401703a43f8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.676898] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006929e6-33dc-4bd4-b8be-ed7574db6b6a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.722764] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe48f47-979d-4548-9c73-7f41ba0d1459 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.731749] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7644aed-4659-449a-b7ea-0c6bca62e0cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.752698] env[65107]: DEBUG nova.compute.provider_tree [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.762973] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 897.842865] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102978, 'name': Rename_Task, 'duration_secs': 0.434022} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.846168] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.846847] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91c69ca7-bb03-4f28-93bd-339c3e7ca58f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.857146] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102979, 'name': ReconfigVM_Task, 'duration_secs': 0.318907} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.858635] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Reconfigured VM instance instance-00000044 to attach disk [datastore1] ac631507-8abf-4019-bdd5-7684dd51d1e9/ac631507-8abf-4019-bdd5-7684dd51d1e9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.859420] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 897.859420] env[65107]: value = "task-5102980" [ 897.859420] env[65107]: _type = "Task" [ 897.859420] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.859620] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94f6ce01-b15e-4960-a6e9-c680982ac1fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.874873] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102980, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.876752] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 897.876752] env[65107]: value = "task-5102981" [ 897.876752] env[65107]: _type = "Task" [ 897.876752] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.886331] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102981, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.908491] env[65107]: WARNING neutronclient.v2_0.client [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.908491] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.908491] env[65107]: WARNING openstack [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.948519] env[65107]: DEBUG nova.network.neutron [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Updating instance_info_cache with network_info: [{"id": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "address": "fa:16:3e:d4:a3:9f", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4659264c-96", "ovs_interfaceid": "4659264c-9645-4cc9-a6aa-4d3299ae4c31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 897.982832] env[65107]: DEBUG nova.network.neutron [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.185504] env[65107]: DEBUG nova.network.neutron [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updated VIF entry in instance network info cache for port b3279b33-9092-457f-88cc-4c2cc95e6cdc. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 898.185985] env[65107]: DEBUG nova.network.neutron [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updating instance_info_cache with network_info: [{"id": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "address": "fa:16:3e:18:a3:54", "network": {"id": "39ea7370-9401-4188-a7d0-a9b311fc29aa", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-53213752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7db97cf40794c5196e31ab5fa8a0477", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3279b33-90", "ovs_interfaceid": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.229692] env[65107]: DEBUG nova.compute.manager [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Received event network-vif-plugged-355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 898.229960] env[65107]: DEBUG oslo_concurrency.lockutils [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Acquiring lock "ac631507-8abf-4019-bdd5-7684dd51d1e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.232513] env[65107]: DEBUG oslo_concurrency.lockutils [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.232761] env[65107]: DEBUG oslo_concurrency.lockutils [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.232883] env[65107]: DEBUG nova.compute.manager [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] No waiting events found dispatching network-vif-plugged-355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 898.233076] env[65107]: WARNING nova.compute.manager [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Received unexpected event network-vif-plugged-355ade04-dfe9-489a-81d1-b7bee4c4833e for instance with vm_state building and task_state spawning. [ 898.233317] env[65107]: DEBUG nova.compute.manager [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Received event network-changed-355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 898.233584] env[65107]: DEBUG nova.compute.manager [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Refreshing instance network info cache due to event network-changed-355ade04-dfe9-489a-81d1-b7bee4c4833e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 898.233725] env[65107]: DEBUG oslo_concurrency.lockutils [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Acquiring lock "refresh_cache-ac631507-8abf-4019-bdd5-7684dd51d1e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.233873] env[65107]: DEBUG oslo_concurrency.lockutils [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Acquired lock "refresh_cache-ac631507-8abf-4019-bdd5-7684dd51d1e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.234038] env[65107]: DEBUG nova.network.neutron [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Refreshing network info cache for port 355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 898.236887] env[65107]: DEBUG nova.network.neutron [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Successfully updated port: 03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 898.276153] env[65107]: ERROR nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] [req-3d8d4c3a-2bbd-4e78-8568-3d3dabc05d8f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3d8d4c3a-2bbd-4e78-8568-3d3dabc05d8f"}]} [ 898.290344] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.292179] env[65107]: DEBUG nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 898.309760] env[65107]: DEBUG nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 898.310162] env[65107]: DEBUG nova.compute.provider_tree [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.323566] env[65107]: DEBUG nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 898.346495] env[65107]: DEBUG nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 898.379944] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102980, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.395918] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102981, 'name': Rename_Task, 'duration_secs': 0.154104} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.396459] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.396816] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f3493a9-7cad-4703-b0ca-6d9a695fbcef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.409507] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 898.409507] env[65107]: value = "task-5102982" [ 898.409507] env[65107]: _type = "Task" [ 898.409507] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.418609] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.451927] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-40e17da0-b459-4619-8e9e-18c3093328c4" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.485954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.689784] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b76a300-7b3d-471a-a9c8-40513888f128 req-b882e9b3-5177-415d-88ff-9ba35c5f4dc9 service nova] Releasing lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.741265] env[65107]: WARNING neutronclient.v2_0.client [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 898.741265] env[65107]: WARNING openstack [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.741627] env[65107]: WARNING openstack [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.751902] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.752119] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquired lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.752338] env[65107]: DEBUG nova.network.neutron [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 898.850512] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378eab48-6bb6-4200-ae66-d484a7158dec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.860591] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e146cffc-385f-4955-86a5-5bb33be45a78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.873115] env[65107]: WARNING openstack [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.873560] env[65107]: WARNING openstack [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.916484] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6b2cf2-b02f-403e-b758-1ecd04809398 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.922504] env[65107]: DEBUG oslo_vmware.api [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5102980, 'name': PowerOnVM_Task, 'duration_secs': 0.722579} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.925424] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.925424] env[65107]: INFO nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Took 13.28 seconds to spawn the instance on the hypervisor. [ 898.925424] env[65107]: DEBUG nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 898.925424] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0fd605-804f-4af9-9826-1c6ce2368c4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.932878] env[65107]: DEBUG oslo_vmware.api [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102982, 'name': PowerOnVM_Task, 'duration_secs': 0.492829} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.934714] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4644fce-378f-49eb-b323-23f73d4c4029 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.939477] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.939720] env[65107]: INFO nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Took 10.67 seconds to spawn the instance on the hypervisor. [ 898.939906] env[65107]: DEBUG nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 898.942460] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050f7b2a-4ffb-429e-bf4a-d8851ace2001 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.965938] env[65107]: DEBUG nova.compute.provider_tree [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.032503] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52315503-d7d9-410e-8aa7-a544d1786341 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.036481] env[65107]: WARNING neutronclient.v2_0.client [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.037059] env[65107]: WARNING openstack [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.037415] env[65107]: WARNING openstack [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.063912] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b955f15-8189-4d67-9f5c-a38e61f4f173 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.074904] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 899.141366] env[65107]: DEBUG nova.network.neutron [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Updated VIF entry in instance network info cache for port 355ade04-dfe9-489a-81d1-b7bee4c4833e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 899.141743] env[65107]: DEBUG nova.network.neutron [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Updating instance_info_cache with network_info: [{"id": "355ade04-dfe9-489a-81d1-b7bee4c4833e", "address": "fa:16:3e:de:1f:ee", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap355ade04-df", "ovs_interfaceid": "355ade04-dfe9-489a-81d1-b7bee4c4833e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 899.235386] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.235869] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.236119] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.236119] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.236331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.238370] env[65107]: INFO nova.compute.manager [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Terminating instance [ 899.260025] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.260025] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.296280] env[65107]: DEBUG nova.network.neutron [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 899.323383] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.323383] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.394925] env[65107]: WARNING neutronclient.v2_0.client [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.395597] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.395938] env[65107]: WARNING openstack [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.464165] env[65107]: INFO nova.compute.manager [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Took 35.07 seconds to build instance. [ 899.483032] env[65107]: INFO nova.compute.manager [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Took 34.74 seconds to build instance. [ 899.512979] env[65107]: DEBUG nova.network.neutron [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Updating instance_info_cache with network_info: [{"id": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "address": "fa:16:3e:78:ab:3f", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03fa2dfb-b3", "ovs_interfaceid": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 899.538850] env[65107]: DEBUG nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 899.539184] env[65107]: DEBUG nova.compute.provider_tree [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 97 to 98 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 899.539421] env[65107]: DEBUG nova.compute.provider_tree [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.581099] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b87e25ed-9133-44c3-9222-cd99373347b8 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance '71e6e280-50c3-4655-a6a3-40ea4c650d96' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 899.645081] env[65107]: DEBUG oslo_concurrency.lockutils [req-acdcf7c5-5858-4809-ae48-f5b48131f7be req-0a5e916d-2a53-456f-9c58-60765cd4703a service nova] Releasing lock "refresh_cache-ac631507-8abf-4019-bdd5-7684dd51d1e9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 899.743480] env[65107]: DEBUG nova.compute.manager [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 899.743480] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.744325] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10339449-b1c5-4dce-853c-c9f76f0807a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.753170] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.753871] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab6fa3b2-6839-4702-813d-e1633c7c2702 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.760062] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 899.760062] env[65107]: value = "task-5102983" [ 899.760062] env[65107]: _type = "Task" [ 899.760062] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.771297] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102983, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.796332] env[65107]: DEBUG nova.compute.manager [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Received event network-changed-190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 899.796332] env[65107]: DEBUG nova.compute.manager [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Refreshing instance network info cache due to event network-changed-190c1749-3fb8-4199-b8e7-1ab227d2b30f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 899.796332] env[65107]: DEBUG oslo_concurrency.lockutils [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Acquiring lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.798405] env[65107]: DEBUG oslo_concurrency.lockutils [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Acquired lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.798621] env[65107]: DEBUG nova.network.neutron [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Refreshing network info cache for port 190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 899.971111] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6d8eab45-c676-471e-956c-2d4fe99b39de tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.585s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.988265] env[65107]: DEBUG oslo_concurrency.lockutils [None req-de0d9950-4995-4978-af62-464a4ac42ae5 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.262s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.009295] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "fead9fe9-23f4-4d13-bf44-be0727057ddd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.009988] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.019679] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Releasing lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 900.020245] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Instance network_info: |[{"id": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "address": "fa:16:3e:78:ab:3f", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03fa2dfb-b3", "ovs_interfaceid": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 900.020862] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.021457] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:ab:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03fa2dfb-b370-41bb-ab18-1d3af50fd4d9', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.029721] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Creating folder: Project (cd4b1120670a4df9af1fb8905c4b2b8f). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 900.030806] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c91d77ef-13f9-4dcb-97cc-5066d4e7495f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.033851] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffa4f2aa-753e-47f1-8269-eba2e9325d54 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.044206] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 900.044206] env[65107]: value = "task-5102984" [ 900.044206] env[65107]: _type = "Task" [ 900.044206] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.052437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.355s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.055601] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Created folder: Project (cd4b1120670a4df9af1fb8905c4b2b8f) in parent group-v992574. [ 900.055953] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Creating folder: Instances. Parent ref: group-v992771. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 900.056735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.490s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.057154] env[65107]: DEBUG nova.objects.instance [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'resources' on Instance uuid 701a1a16-9348-4748-bb24-ee245a76566c {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.062634] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16d40d53-6df0-44b2-947f-b66eea6b6df2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.070265] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102984, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.082669] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Created folder: Instances in parent group-v992771. [ 900.082669] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 900.082669] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.082669] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35dc1d8c-396b-4817-936c-cad062a37aa4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.105213] env[65107]: INFO nova.scheduler.client.report [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Deleted allocations for instance e3510c31-d6be-4e9f-a0a2-a662123861e9 [ 900.121507] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.121507] env[65107]: value = "task-5102987" [ 900.121507] env[65107]: _type = "Task" [ 900.121507] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.132811] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102987, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.272353] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102983, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.302703] env[65107]: WARNING neutronclient.v2_0.client [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 900.303484] env[65107]: WARNING openstack [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.303861] env[65107]: WARNING openstack [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.453174] env[65107]: WARNING openstack [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.453543] env[65107]: WARNING openstack [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.460255] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.460481] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.511844] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 900.536042] env[65107]: WARNING neutronclient.v2_0.client [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 900.536713] env[65107]: WARNING openstack [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.537094] env[65107]: WARNING openstack [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.555661] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102984, 'name': PowerOffVM_Task, 'duration_secs': 0.371049} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.555661] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.556579] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5bc507-a58b-465c-8c59-586404375067 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.575959] env[65107]: DEBUG nova.objects.instance [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'numa_topology' on Instance uuid 701a1a16-9348-4748-bb24-ee245a76566c {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.578680] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e445350-a921-429b-8559-334ec904ba20 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.617208] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.619767] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ad946f5-4bdb-4586-b4c3-87d50bd0087c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.622562] env[65107]: DEBUG oslo_concurrency.lockutils [None req-705dc0c3-d08f-47de-8aa0-1f6fab0207c1 tempest-ServersAdminTestJSON-1629823880 tempest-ServersAdminTestJSON-1629823880-project-member] Lock "e3510c31-d6be-4e9f-a0a2-a662123861e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.705s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.634974] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102987, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.637484] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 900.637484] env[65107]: value = "task-5102988" [ 900.637484] env[65107]: _type = "Task" [ 900.637484] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.640947] env[65107]: DEBUG nova.network.neutron [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updated VIF entry in instance network info cache for port 190c1749-3fb8-4199-b8e7-1ab227d2b30f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 900.641376] env[65107]: DEBUG nova.network.neutron [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updating instance_info_cache with network_info: [{"id": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "address": "fa:16:3e:39:cf:6d", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap190c1749-3f", "ovs_interfaceid": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 900.655573] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 900.655824] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.656120] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.656297] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 900.656538] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.657087] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01e295a0-69fc-4cae-a451-a84d530c0f5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.673571] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.673571] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.674250] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19d8b5ed-0819-458f-b91e-ec79fc655aec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.682047] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 900.682047] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8f45-63a9-1f30-7e30-c87e6dfb281b" [ 900.682047] env[65107]: _type = "Task" [ 900.682047] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.699604] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8f45-63a9-1f30-7e30-c87e6dfb281b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.772880] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102983, 'name': PowerOffVM_Task, 'duration_secs': 0.602984} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.774139] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.774139] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.774139] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-892e5158-a8d9-4032-9ced-f53c7e30087e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.964580] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 901.029507] env[65107]: DEBUG nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Received event network-changed-875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 901.030020] env[65107]: DEBUG nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Refreshing instance network info cache due to event network-changed-875128de-2b86-4f3c-bef3-92733849587c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 901.030020] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Acquiring lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.030317] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Acquired lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.030430] env[65107]: DEBUG nova.network.neutron [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Refreshing network info cache for port 875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 901.043247] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.071160] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.071485] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.071580] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Deleting the datastore file [datastore2] 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.074059] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d25e59b-7a35-4112-b543-4f094f29d166 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.080941] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for the task: (returnval){ [ 901.080941] env[65107]: value = "task-5102990" [ 901.080941] env[65107]: _type = "Task" [ 901.080941] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.085403] env[65107]: DEBUG nova.objects.base [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Object Instance<701a1a16-9348-4748-bb24-ee245a76566c> lazy-loaded attributes: resources,numa_topology {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 901.097501] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.145569] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5102987, 'name': CreateVM_Task, 'duration_secs': 0.789193} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.145825] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.146722] env[65107]: WARNING neutronclient.v2_0.client [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.146854] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.146958] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.147343] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 901.150665] env[65107]: DEBUG oslo_concurrency.lockutils [req-58c2d836-ff8f-423a-a1f4-41b11584fd63 req-1597bea5-b2a2-418c-8fa5-b05e47dd27eb service nova] Releasing lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.151025] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53927667-daaf-4b73-a9d8-adbcc6daf089 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.157965] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 901.157965] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ece9e9-9dda-b2d1-1165-7530b5c8cd51" [ 901.157965] env[65107]: _type = "Task" [ 901.157965] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.170722] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ece9e9-9dda-b2d1-1165-7530b5c8cd51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.195356] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8f45-63a9-1f30-7e30-c87e6dfb281b, 'name': SearchDatastore_Task, 'duration_secs': 0.029655} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.199419] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7c0dbe0-950e-43fe-a2c4-f7b6e98d0a7c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.203416] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.204053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.204053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.204053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.204194] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.206386] env[65107]: INFO nova.compute.manager [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Terminating instance [ 901.212624] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 901.212624] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528c4950-0cd3-2221-045e-657829ca94f8" [ 901.212624] env[65107]: _type = "Task" [ 901.212624] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.229303] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528c4950-0cd3-2221-045e-657829ca94f8, 'name': SearchDatastore_Task, 'duration_secs': 0.013032} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.229565] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.229824] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. {{(pid=65107) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 901.230422] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a04b7f2-471f-4a98-a338-a7367a469c60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.242523] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 901.242523] env[65107]: value = "task-5102991" [ 901.242523] env[65107]: _type = "Task" [ 901.242523] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.254783] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.328085] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.328367] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.328619] env[65107]: INFO nova.compute.manager [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Rebooting instance [ 901.486730] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.534506] env[65107]: WARNING neutronclient.v2_0.client [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.535644] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.536581] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.564438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f00f26-fac4-4a38-8109-d011492a810b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.575409] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e4fe1a-797e-4358-86c4-cbd00563c32e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.613591] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc952696-f200-4c10-8682-0bd42ff772bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.619917] env[65107]: DEBUG oslo_vmware.api [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Task: {'id': task-5102990, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248465} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.620545] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.620798] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.621012] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.621231] env[65107]: INFO nova.compute.manager [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Took 1.88 seconds to destroy the instance on the hypervisor. [ 901.621507] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 901.621749] env[65107]: DEBUG nova.compute.manager [-] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 901.621868] env[65107]: DEBUG nova.network.neutron [-] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 901.622144] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.622757] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.622972] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.635761] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d196069-add6-4966-a26b-5345e43e813d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.655590] env[65107]: DEBUG nova.compute.provider_tree [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.670848] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ece9e9-9dda-b2d1-1165-7530b5c8cd51, 'name': SearchDatastore_Task, 'duration_secs': 0.022981} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.671986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.673026] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.673026] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.673026] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.673026] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.673693] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08e6ba33-1840-4b42-92d4-f3ee93271469 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.683058] env[65107]: DEBUG nova.compute.manager [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 901.683325] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ddd02f-f57b-469c-beab-66cb02d793e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.689151] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.689618] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.692907] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64c457ce-4290-407a-be54-6c547ad4f179 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.703200] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 901.703200] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e47c35-4a73-e916-46fb-de3fb831f228" [ 901.703200] env[65107]: _type = "Task" [ 901.703200] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.714396] env[65107]: DEBUG nova.compute.manager [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 901.714396] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.714711] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e47c35-4a73-e916-46fb-de3fb831f228, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.718030] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae992149-7ad8-4945-90e9-57292f36fa61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.727462] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.727528] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4715162c-98ef-4e7d-80ba-ce1000140b38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.736776] env[65107]: DEBUG oslo_vmware.api [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 901.736776] env[65107]: value = "task-5102992" [ 901.736776] env[65107]: _type = "Task" [ 901.736776] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.750979] env[65107]: DEBUG oslo_vmware.api [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.760506] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102991, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.848572] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.858896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.859113] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquired lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.859293] env[65107]: DEBUG nova.network.neutron [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 902.162588] env[65107]: DEBUG nova.scheduler.client.report [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 902.202223] env[65107]: INFO nova.compute.manager [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] instance snapshotting [ 902.209691] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d733389-a764-4456-b57a-045a557fb08c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.222290] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e47c35-4a73-e916-46fb-de3fb831f228, 'name': SearchDatastore_Task, 'duration_secs': 0.040389} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.237750] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10bc9f27-1587-49fd-8af5-52df0247f956 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.242956] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc2a594-55f5-432e-ab2b-cd61a15095a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.261335] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 902.261335] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8077e-cee4-6d20-42a6-a2cfcebf1460" [ 902.261335] env[65107]: _type = "Task" [ 902.261335] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.278883] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610494} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.280097] env[65107]: DEBUG oslo_vmware.api [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102992, 'name': PowerOffVM_Task, 'duration_secs': 0.311783} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.280097] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. [ 902.280381] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.283121] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.283121] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5934a553-831a-43ec-89a2-02d519ac018d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.290820] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9d0226a-8431-426d-a079-e3c8d650de74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.292202] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f8077e-cee4-6d20-42a6-a2cfcebf1460, 'name': SearchDatastore_Task, 'duration_secs': 0.020718} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.293215] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.293516] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 1a813da6-f777-4888-b2ae-fbad58e01f61/1a813da6-f777-4888-b2ae-fbad58e01f61.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.293816] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50d5ee1d-553d-4d8c-8b2a-521bd1e0a475 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.318925] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.320772] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.321508] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.329873] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b24c150-e111-4908-91ee-e20c5ec05d73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.345944] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 902.345944] env[65107]: value = "task-5102994" [ 902.345944] env[65107]: _type = "Task" [ 902.345944] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.351653] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 902.351653] env[65107]: value = "task-5102995" [ 902.351653] env[65107]: _type = "Task" [ 902.351653] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.359296] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5102994, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.365262] env[65107]: WARNING neutronclient.v2_0.client [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 902.367295] env[65107]: WARNING openstack [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.367295] env[65107]: WARNING openstack [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.374842] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102995, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.401725] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.401725] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.401725] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Deleting the datastore file [datastore2] fb4229c6-b80d-49e9-a68b-baa12c0004ed {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.401961] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfad4cc2-69aa-4b6e-be01-4bc1ba96fa02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.410254] env[65107]: DEBUG oslo_vmware.api [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for the task: (returnval){ [ 902.410254] env[65107]: value = "task-5102996" [ 902.410254] env[65107]: _type = "Task" [ 902.410254] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.420269] env[65107]: DEBUG oslo_vmware.api [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.589288] env[65107]: WARNING neutronclient.v2_0.client [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 902.590068] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.590536] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.669499] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.613s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 902.676021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 21.105s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.676021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 902.676021] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 902.676021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.518s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.676021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 902.680644] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.854s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.682331] env[65107]: INFO nova.compute.claims [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.697474] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05e79d0-10e7-4373-b098-6ddeb04d8011 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.716025] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5565f732-b1ae-442b-831a-0a67426f8c4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.735941] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb0b367-22ff-4479-b85f-134654a459e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.750747] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb04ee9-36da-48dd-86c4-e14c04140c57 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.765756] env[65107]: INFO nova.scheduler.client.report [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleted allocations for instance 369caf24-48a6-4640-8601-e30fd47453da [ 902.800295] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 902.800820] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178583MB free_disk=94GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 902.800974] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 902.804557] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-328bdadc-e67d-4684-9d73-959253f2c1fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.815841] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 902.815841] env[65107]: value = "task-5102997" [ 902.815841] env[65107]: _type = "Task" [ 902.815841] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.829754] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102997, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.831169] env[65107]: WARNING openstack [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.831762] env[65107]: WARNING openstack [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.859972] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5102994, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.867505] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102995, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.874373] env[65107]: DEBUG nova.network.neutron [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Updated VIF entry in instance network info cache for port 875128de-2b86-4f3c-bef3-92733849587c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 902.874373] env[65107]: DEBUG nova.network.neutron [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Updating instance_info_cache with network_info: [{"id": "875128de-2b86-4f3c-bef3-92733849587c", "address": "fa:16:3e:55:fd:2a", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap875128de-2b", "ovs_interfaceid": "875128de-2b86-4f3c-bef3-92733849587c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 902.924020] env[65107]: DEBUG oslo_vmware.api [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Task: {'id': task-5102996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206015} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.924020] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.924020] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.924020] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.924020] env[65107]: INFO nova.compute.manager [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Took 1.21 seconds to destroy the instance on the hypervisor. [ 902.924020] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 902.924375] env[65107]: DEBUG nova.compute.manager [-] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 902.924375] env[65107]: DEBUG nova.network.neutron [-] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 902.924524] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 902.925345] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.925741] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.001079] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.023466] env[65107]: DEBUG nova.network.neutron [-] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.031804] env[65107]: WARNING neutronclient.v2_0.client [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.032653] env[65107]: WARNING openstack [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.033018] env[65107]: WARNING openstack [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.195222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-db2effc1-28b5-43c5-87fd-1496415d2140 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 49.863s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.196416] env[65107]: DEBUG oslo_concurrency.lockutils [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 23.821s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.196731] env[65107]: DEBUG oslo_concurrency.lockutils [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "701a1a16-9348-4748-bb24-ee245a76566c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.196948] env[65107]: DEBUG oslo_concurrency.lockutils [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.197129] env[65107]: DEBUG oslo_concurrency.lockutils [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.198750] env[65107]: INFO nova.compute.manager [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Terminating instance [ 903.310450] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2e87e5b-e002-4919-8b08-687745093a38 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "369caf24-48a6-4640-8601-e30fd47453da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.328s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.331558] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102997, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.361976] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5102994, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568952} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.363028] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 1a813da6-f777-4888-b2ae-fbad58e01f61/1a813da6-f777-4888-b2ae-fbad58e01f61.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.363643] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.363643] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11571d2b-4ea9-4322-88b3-b16f807bbb8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.368546] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102995, 'name': ReconfigVM_Task, 'duration_secs': 0.68736} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.369251] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.370076] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c3fae0-5cb4-4ae9-96ca-c3e7a8d09054 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.375425] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 903.375425] env[65107]: value = "task-5102998" [ 903.375425] env[65107]: _type = "Task" [ 903.375425] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.397162] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Releasing lock "refresh_cache-3f34d6a2-0370-4239-9556-b4333e454d58" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.398037] env[65107]: DEBUG nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Received event network-vif-plugged-03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 903.398037] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Acquiring lock "1a813da6-f777-4888-b2ae-fbad58e01f61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.398037] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.398209] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.398311] env[65107]: DEBUG nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] No waiting events found dispatching network-vif-plugged-03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 903.398471] env[65107]: WARNING nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Received unexpected event network-vif-plugged-03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 for instance with vm_state building and task_state spawning. [ 903.398626] env[65107]: DEBUG nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Received event network-changed-03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 903.398773] env[65107]: DEBUG nova.compute.manager [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Refreshing instance network info cache due to event network-changed-03fa2dfb-b370-41bb-ab18-1d3af50fd4d9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 903.398947] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Acquiring lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.399109] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Acquired lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.399274] env[65107]: DEBUG nova.network.neutron [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Refreshing network info cache for port 03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 903.409320] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86862073-b3e9-473e-8fb2-0f823c745669 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.426051] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5102998, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.428941] env[65107]: DEBUG nova.network.neutron [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updating instance_info_cache with network_info: [{"id": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "address": "fa:16:3e:39:cf:6d", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap190c1749-3f", "ovs_interfaceid": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.438013] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 903.438013] env[65107]: value = "task-5102999" [ 903.438013] env[65107]: _type = "Task" [ 903.438013] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.449602] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102999, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.527197] env[65107]: INFO nova.compute.manager [-] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Took 1.91 seconds to deallocate network for instance. [ 903.574256] env[65107]: DEBUG nova.compute.manager [req-21bfb53a-1fe9-4bc3-aacd-4df9caecf39c req-1adec71f-0659-41d6-af7a-598a2b92c026 service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received event network-vif-deleted-c2fdc887-a837-4102-9f3a-85fec34ad2b9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 903.703218] env[65107]: DEBUG nova.compute.manager [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 903.703613] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.707916] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6918ed7e-f7d8-4ba2-949d-067efbb4294a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.719226] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e7dc42-73dd-4e06-8286-ddf1d9f6c3dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.759149] env[65107]: WARNING nova.virt.vmwareapi.vmops [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 701a1a16-9348-4748-bb24-ee245a76566c could not be found. [ 903.759149] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.759279] env[65107]: INFO nova.compute.manager [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Took 0.06 seconds to destroy the instance on the hypervisor. [ 903.759560] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 903.762689] env[65107]: DEBUG nova.compute.manager [-] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 903.762689] env[65107]: DEBUG nova.network.neutron [-] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 903.762864] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.763404] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.763665] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.796522] env[65107]: DEBUG nova.compute.manager [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Received event network-changed-b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 903.796522] env[65107]: DEBUG nova.compute.manager [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Refreshing instance network info cache due to event network-changed-b3279b33-9092-457f-88cc-4c2cc95e6cdc. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 903.797107] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Acquiring lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.797367] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Acquired lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.797581] env[65107]: DEBUG nova.network.neutron [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Refreshing network info cache for port b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 903.812523] env[65107]: DEBUG nova.network.neutron [-] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.835043] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102997, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.890200] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5102998, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069157} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.891430] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.893643] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.895819] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e73329-f567-4551-91a7-510fd07d2950 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.912896] env[65107]: WARNING neutronclient.v2_0.client [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 903.915278] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.916151] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.933846] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 1a813da6-f777-4888-b2ae-fbad58e01f61/1a813da6-f777-4888-b2ae-fbad58e01f61.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.938166] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Releasing lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.938377] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9afc8b52-3966-4f89-b027-ca05b3ab8020 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.964614] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5102999, 'name': ReconfigVM_Task, 'duration_secs': 0.373612} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.965768] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.966508] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 903.966508] env[65107]: value = "task-5103000" [ 903.966508] env[65107]: _type = "Task" [ 903.966508] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.966639] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbda2814-3dfb-4c0d-bb0f-ec6a747b9fd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.980511] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103000, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.982076] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 903.982076] env[65107]: value = "task-5103001" [ 903.982076] env[65107]: _type = "Task" [ 903.982076] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.994826] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.036902] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.234136] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fd32ac-ecfe-4085-abd4-c6489a1535b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.243568] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b724cb-a27c-438b-87de-06b8966aa810 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.281159] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633fbeaa-9df5-4db5-a5fe-8afe9d243447 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.292552] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed1af29-722b-4989-a493-29428dbbd973 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.307489] env[65107]: DEBUG nova.compute.provider_tree [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.310917] env[65107]: WARNING neutronclient.v2_0.client [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.311636] env[65107]: WARNING openstack [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.311988] env[65107]: WARNING openstack [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.320715] env[65107]: INFO nova.compute.manager [-] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Took 1.40 seconds to deallocate network for instance. [ 904.335837] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.336123] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.336313] env[65107]: DEBUG nova.compute.manager [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Going to confirm migration 3 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 904.337756] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5102997, 'name': CreateSnapshot_Task, 'duration_secs': 1.09823} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.338340] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 904.339399] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af81227-4f30-4fbb-8e39-ad7bd0b6ee0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.444108] env[65107]: DEBUG nova.compute.manager [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 904.445060] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9173e691-203d-40eb-955f-c5d53af06712 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.460997] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.461417] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.478427] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103000, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.492118] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103001, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.613182] env[65107]: WARNING openstack [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.613182] env[65107]: WARNING openstack [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.642081] env[65107]: WARNING neutronclient.v2_0.client [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.642990] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.643375] env[65107]: WARNING openstack [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.654232] env[65107]: DEBUG nova.network.neutron [-] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 904.748335] env[65107]: WARNING neutronclient.v2_0.client [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.748335] env[65107]: WARNING openstack [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.748639] env[65107]: WARNING openstack [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.792097] env[65107]: DEBUG nova.network.neutron [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Updated VIF entry in instance network info cache for port 03fa2dfb-b370-41bb-ab18-1d3af50fd4d9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 904.792492] env[65107]: DEBUG nova.network.neutron [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Updating instance_info_cache with network_info: [{"id": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "address": "fa:16:3e:78:ab:3f", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03fa2dfb-b3", "ovs_interfaceid": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 904.810793] env[65107]: DEBUG nova.scheduler.client.report [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 904.830954] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.843199] env[65107]: WARNING neutronclient.v2_0.client [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.864350] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 904.870268] env[65107]: DEBUG nova.network.neutron [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updated VIF entry in instance network info cache for port b3279b33-9092-457f-88cc-4c2cc95e6cdc. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 904.870695] env[65107]: DEBUG nova.network.neutron [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updating instance_info_cache with network_info: [{"id": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "address": "fa:16:3e:18:a3:54", "network": {"id": "39ea7370-9401-4188-a7d0-a9b311fc29aa", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-53213752-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7db97cf40794c5196e31ab5fa8a0477", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3279b33-90", "ovs_interfaceid": "b3279b33-9092-457f-88cc-4c2cc95e6cdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 904.872128] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-16ca737e-43d6-4764-bcb7-18d3c07a7b33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.881883] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 904.881883] env[65107]: value = "task-5103002" [ 904.881883] env[65107]: _type = "Task" [ 904.881883] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.892474] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103002, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.912628] env[65107]: WARNING neutronclient.v2_0.client [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.913025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.913204] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.913582] env[65107]: DEBUG nova.network.neutron [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 904.913824] env[65107]: DEBUG nova.objects.instance [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'info_cache' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.979457] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103000, 'name': ReconfigVM_Task, 'duration_secs': 0.754874} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.979789] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 1a813da6-f777-4888-b2ae-fbad58e01f61/1a813da6-f777-4888-b2ae-fbad58e01f61.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.981559] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed6d465a-3377-4c1b-93d0-3fe02a168cb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.995049] env[65107]: DEBUG oslo_vmware.api [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103001, 'name': PowerOnVM_Task, 'duration_secs': 0.712102} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.996610] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.998421] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 904.998421] env[65107]: value = "task-5103003" [ 904.998421] env[65107]: _type = "Task" [ 904.998421] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.999466] env[65107]: DEBUG nova.compute.manager [None req-d86c567a-1147-4df7-a3da-77bde4dc9348 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 905.004109] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c4da0e-dee6-4d4f-bc50-acafff52a4a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.018269] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103003, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.156909] env[65107]: INFO nova.compute.manager [-] [instance: 701a1a16-9348-4748-bb24-ee245a76566c] Took 1.39 seconds to deallocate network for instance. [ 905.259038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.259038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.259038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.259038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.259038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.263170] env[65107]: INFO nova.compute.manager [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Terminating instance [ 905.296187] env[65107]: DEBUG oslo_concurrency.lockutils [req-348c8e4c-3a03-4fd9-88d2-ab01c739fa9a req-a9a9234e-d8e8-4c73-b9c0-f99544213c89 service nova] Releasing lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.329952] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.329952] env[65107]: DEBUG nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 905.333534] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.744s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.336458] env[65107]: INFO nova.compute.claims [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.376615] env[65107]: DEBUG oslo_concurrency.lockutils [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] Releasing lock "refresh_cache-cddb14db-34c3-4783-b21e-e0b4623a3a9e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.376880] env[65107]: DEBUG nova.compute.manager [req-6b7a7d80-dcca-4029-a688-b771c7644d5d req-c041a5d6-1e14-4f9e-9a59-31b68b274365 service nova] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Received event network-vif-deleted-33f13ff6-7d02-4931-bda5-04c0658cfd6c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 905.398641] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103002, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.465989] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e39ad9-8b06-4558-a1ae-e74b40c83064 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.474570] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Doing hard reboot of VM {{(pid=65107) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 905.474856] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0401b383-1410-4467-9324-01fa888cea53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.483247] env[65107]: DEBUG oslo_vmware.api [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 905.483247] env[65107]: value = "task-5103004" [ 905.483247] env[65107]: _type = "Task" [ 905.483247] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.492864] env[65107]: DEBUG oslo_vmware.api [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103004, 'name': ResetVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.515067] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103003, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.722263] env[65107]: DEBUG nova.compute.manager [req-5452797f-4ab0-463c-96cb-bb22e5016242 req-f2d7c393-a890-488d-a258-dd1f1a5cae72 service nova] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Received event network-vif-deleted-c23ac068-c418-45e9-8edc-0e541664cbef {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 905.769640] env[65107]: DEBUG nova.compute.manager [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 905.769640] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 905.769640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384f07fe-ff7f-437a-becf-0029f6b6713a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.779296] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.779622] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f72a662-b610-4e7d-bb96-7b9ca285787f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.787469] env[65107]: DEBUG oslo_vmware.api [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 905.787469] env[65107]: value = "task-5103005" [ 905.787469] env[65107]: _type = "Task" [ 905.787469] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.798730] env[65107]: DEBUG oslo_vmware.api [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5103005, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.843702] env[65107]: DEBUG nova.compute.utils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 905.857729] env[65107]: DEBUG nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 905.897770] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103002, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.921152] env[65107]: WARNING neutronclient.v2_0.client [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 905.921778] env[65107]: WARNING openstack [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.922077] env[65107]: WARNING openstack [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.995662] env[65107]: DEBUG oslo_vmware.api [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103004, 'name': ResetVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.014016] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103003, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.186566] env[65107]: DEBUG oslo_concurrency.lockutils [None req-25ae0973-f3ac-4589-bf42-6d8520645b36 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "701a1a16-9348-4748-bb24-ee245a76566c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.990s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.240375] env[65107]: WARNING openstack [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.241251] env[65107]: WARNING openstack [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.300034] env[65107]: DEBUG oslo_vmware.api [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5103005, 'name': PowerOffVM_Task, 'duration_secs': 0.227326} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.300434] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.302126] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.302126] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3d1c436-1b7c-41f5-b7e4-c17d9f821c92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.356460] env[65107]: DEBUG nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 906.364441] env[65107]: WARNING neutronclient.v2_0.client [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.365258] env[65107]: WARNING openstack [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.366318] env[65107]: WARNING openstack [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.383241] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.383472] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.383656] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleting the datastore file [datastore2] fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.383990] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6697c69-60ed-430d-ba6a-db29eac84d14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.400208] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103002, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.406262] env[65107]: DEBUG oslo_vmware.api [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for the task: (returnval){ [ 906.406262] env[65107]: value = "task-5103007" [ 906.406262] env[65107]: _type = "Task" [ 906.406262] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.421156] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "88cac920-3c1e-41dc-9e52-f51f48a80f56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.421619] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.421863] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "88cac920-3c1e-41dc-9e52-f51f48a80f56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.422158] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.423285] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.425107] env[65107]: INFO nova.compute.manager [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Terminating instance [ 906.430208] env[65107]: DEBUG oslo_vmware.api [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5103007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.497159] env[65107]: DEBUG oslo_vmware.api [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103004, 'name': ResetVM_Task, 'duration_secs': 0.780219} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.497587] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Did hard reboot of VM {{(pid=65107) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 906.498078] env[65107]: DEBUG nova.compute.manager [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 906.499401] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53544b0c-e623-4cfb-899f-e096a5974871 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.521243] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103003, 'name': Rename_Task, 'duration_secs': 1.211122} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.521641] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.522270] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-910c0eba-960e-419a-ba40-585695a08135 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.531984] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 906.531984] env[65107]: value = "task-5103008" [ 906.531984] env[65107]: _type = "Task" [ 906.531984] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.542388] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103008, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.644716] env[65107]: DEBUG nova.network.neutron [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 906.888705] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07879ef5-5d86-4844-8584-6f35c8b29700 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.906523] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd94950-06f2-4392-96db-dafe788bef17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.910931] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103002, 'name': CloneVM_Task, 'duration_secs': 1.811878} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.911460] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Created linked-clone VM from snapshot [ 906.917560] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385d55f4-bfe2-4c9b-899d-bf58094665e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.947292] env[65107]: DEBUG nova.compute.manager [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 906.947522] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.949361] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5129324-c17e-4fd5-819c-b08e330daf87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.952881] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc89535f-6f59-4fae-ac65-4656073508c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.963905] env[65107]: DEBUG oslo_vmware.api [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Task: {'id': task-5103007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.252191} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.964623] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Uploading image 57f71a90-320c-4d88-8671-5ad259d2a173 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 906.968903] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.968903] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 906.968903] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.968903] env[65107]: INFO nova.compute.manager [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Took 1.20 seconds to destroy the instance on the hypervisor. [ 906.969453] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 906.969593] env[65107]: DEBUG nova.compute.manager [-] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 906.969701] env[65107]: DEBUG nova.network.neutron [-] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 906.969937] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.970556] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.970817] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.986025] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.986025] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bc09a4-6fe6-4545-a644-8f1ede651da7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.989085] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76c070bd-1dd3-4837-a544-df59d36eeb1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.007245] env[65107]: DEBUG nova.compute.provider_tree [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.013570] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 907.013570] env[65107]: value = "vm-992775" [ 907.013570] env[65107]: _type = "VirtualMachine" [ 907.013570] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 907.013570] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4cade81c-e976-42f7-9335-83ec82b9f1fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.015952] env[65107]: DEBUG oslo_vmware.api [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 907.015952] env[65107]: value = "task-5103009" [ 907.015952] env[65107]: _type = "Task" [ 907.015952] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.025466] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease: (returnval){ [ 907.025466] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529841ad-7d56-ad49-ddad-66cb7b070347" [ 907.025466] env[65107]: _type = "HttpNfcLease" [ 907.025466] env[65107]: } obtained for exporting VM: (result){ [ 907.025466] env[65107]: value = "vm-992775" [ 907.025466] env[65107]: _type = "VirtualMachine" [ 907.025466] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 907.025839] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the lease: (returnval){ [ 907.025839] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529841ad-7d56-ad49-ddad-66cb7b070347" [ 907.025839] env[65107]: _type = "HttpNfcLease" [ 907.025839] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 907.034762] env[65107]: DEBUG oslo_vmware.api [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5103009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.035100] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43b35a6b-d081-4344-990d-40a453ff6211 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.707s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.053887] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 907.053887] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529841ad-7d56-ad49-ddad-66cb7b070347" [ 907.053887] env[65107]: _type = "HttpNfcLease" [ 907.053887] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 907.057501] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 907.057501] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529841ad-7d56-ad49-ddad-66cb7b070347" [ 907.057501] env[65107]: _type = "HttpNfcLease" [ 907.057501] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 907.058304] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103008, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.059106] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb67762-97d0-40bb-9f40-64d890e8aebd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.063350] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 907.074693] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52096df5-7584-38b4-953e-24ea95723e69/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 907.074907] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52096df5-7584-38b4-953e-24ea95723e69/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 907.151201] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.151201] env[65107]: DEBUG nova.objects.instance [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'migration_context' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.233023] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-59661803-d158-47ca-8c41-2c6131d5e57e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.367136] env[65107]: DEBUG nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 907.417638] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 907.417923] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 907.418116] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 907.418311] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 907.418474] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 907.418648] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 907.418866] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.419035] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 907.419207] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 907.419387] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 907.419848] env[65107]: DEBUG nova.virt.hardware [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 907.422279] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aceb351-7409-42b4-9a67-10234f4106a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.427754] env[65107]: DEBUG nova.network.neutron [-] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 907.437395] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69a721e-0992-445b-bb37-463ee5868f12 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.455316] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.461077] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Creating folder: Project (198e62b5f04942f2b5a15eb2c7553094). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.461790] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74a720a2-895b-4e7d-8b6b-a99f3c6dede5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.474612] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Created folder: Project (198e62b5f04942f2b5a15eb2c7553094) in parent group-v992574. [ 907.475556] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Creating folder: Instances. Parent ref: group-v992776. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.475556] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-baabe543-dc22-4a33-ab61-aeb218111c7a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.495125] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Created folder: Instances in parent group-v992776. [ 907.495469] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 907.495720] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.496025] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a42a961-ba01-4b31-a223-8f81d41acdcb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.510731] env[65107]: DEBUG nova.scheduler.client.report [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.527665] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.527665] env[65107]: value = "task-5103013" [ 907.527665] env[65107]: _type = "Task" [ 907.527665] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.536982] env[65107]: DEBUG oslo_vmware.api [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5103009, 'name': PowerOffVM_Task, 'duration_secs': 0.294084} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.537940] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.538344] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.546439] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09221df5-2b92-4f54-8235-55d98c0962cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.549126] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103013, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.557126] env[65107]: DEBUG oslo_vmware.api [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103008, 'name': PowerOnVM_Task, 'duration_secs': 0.648469} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.557465] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.557770] env[65107]: INFO nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Took 10.99 seconds to spawn the instance on the hypervisor. [ 907.558043] env[65107]: DEBUG nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 907.559060] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f7bb02-cd5c-45ab-98df-00be194e5a91 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.636207] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.636455] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.636687] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Deleting the datastore file [datastore2] 88cac920-3c1e-41dc-9e52-f51f48a80f56 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.636987] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ace3eac9-cbe4-432f-bc9e-62b424b6f83d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.645517] env[65107]: DEBUG oslo_vmware.api [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for the task: (returnval){ [ 907.645517] env[65107]: value = "task-5103015" [ 907.645517] env[65107]: _type = "Task" [ 907.645517] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.654292] env[65107]: DEBUG nova.objects.base [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Object Instance<71e6e280-50c3-4655-a6a3-40ea4c650d96> lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 907.658949] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d988a6ce-65aa-4049-bb62-ab23f874cbc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.663513] env[65107]: DEBUG oslo_vmware.api [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5103015, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.682463] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-161b9647-847f-4cc8-afa5-1e5b4a24f6d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.691626] env[65107]: DEBUG oslo_vmware.api [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 907.691626] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b9c207-90c8-383f-515b-7f493ec5b7d9" [ 907.691626] env[65107]: _type = "Task" [ 907.691626] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.706256] env[65107]: DEBUG oslo_vmware.api [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b9c207-90c8-383f-515b-7f493ec5b7d9, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.706750] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 907.934418] env[65107]: INFO nova.compute.manager [-] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Took 0.96 seconds to deallocate network for instance. [ 908.021462] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.688s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.022093] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 908.027985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.686s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.027985] env[65107]: DEBUG nova.objects.instance [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lazy-loading 'resources' on Instance uuid 9a46ff51-4035-4709-ae6c-0fb5397200ab {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.048289] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103013, 'name': CreateVM_Task, 'duration_secs': 0.395827} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.049412] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.049652] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.050419] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.050609] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 908.051227] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cee43d62-3d3a-4a43-875f-4965ac3609c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.059114] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 908.059114] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527232d8-e3e2-ed79-eb4e-bed881f7ca3e" [ 908.059114] env[65107]: _type = "Task" [ 908.059114] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.078365] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527232d8-e3e2-ed79-eb4e-bed881f7ca3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.084287] env[65107]: INFO nova.compute.manager [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Took 39.23 seconds to build instance. [ 908.158625] env[65107]: DEBUG oslo_vmware.api [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Task: {'id': task-5103015, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174914} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.159228] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.159592] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.159923] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.160274] env[65107]: INFO nova.compute.manager [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Took 1.21 seconds to destroy the instance on the hypervisor. [ 908.160652] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 908.160975] env[65107]: DEBUG nova.compute.manager [-] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 908.161197] env[65107]: DEBUG nova.network.neutron [-] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 908.161647] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.165110] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.165110] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.423023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "888d60d4-bc9c-48af-97f9-434883884e8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 908.423023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "888d60d4-bc9c-48af-97f9-434883884e8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.445253] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 908.467712] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.534408] env[65107]: DEBUG nova.compute.utils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 908.537379] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 908.537379] env[65107]: DEBUG nova.network.neutron [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 908.537379] env[65107]: WARNING neutronclient.v2_0.client [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.537379] env[65107]: WARNING neutronclient.v2_0.client [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.537379] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.537718] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.577066] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527232d8-e3e2-ed79-eb4e-bed881f7ca3e, 'name': SearchDatastore_Task, 'duration_secs': 0.013479} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.581625] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.583145] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.583145] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.583145] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.583145] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.583813] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-783ac0f6-f0cf-4b61-830b-15483f49dc84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.586691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fa5c3023-b677-467e-a9ee-523221a5ff14 tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.747s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.600142] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.600316] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.601373] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed65da2c-845f-4adf-b66a-7ec12c9fd4fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.611190] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 908.611190] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d6e999-7647-eb7a-f338-fbf25c8997be" [ 908.611190] env[65107]: _type = "Task" [ 908.611190] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.621746] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d6e999-7647-eb7a-f338-fbf25c8997be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.901014] env[65107]: DEBUG nova.policy [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4e9500837fe468ba2ab9901b9a7803b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05918ba13682405ba5e5e7ebde13b42b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 908.927017] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 909.049146] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 909.106540] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e97a48-4e51-4039-8e71-45595209cca3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.119277] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba3b58c-978c-4d61-a0f9-2c475358d6de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.130877] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d6e999-7647-eb7a-f338-fbf25c8997be, 'name': SearchDatastore_Task, 'duration_secs': 0.015977} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.132183] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ccd1dc-e8a0-4563-92e4-66e3011a6f33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.171804] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19306bee-3406-48f7-8aa7-82b359670f30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.177249] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 909.177249] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e48ece-5470-8f91-43e4-937d3673c09e" [ 909.177249] env[65107]: _type = "Task" [ 909.177249] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.186008] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f696cafb-5b01-4bea-9ea3-d56a072c6222 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.197613] env[65107]: DEBUG nova.compute.manager [req-fb5043db-1fa9-4197-8320-3230f95eb0fc req-92c67ada-46c9-4cb2-a638-2a8056020a30 service nova] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Received event network-vif-deleted-086361c0-3cd7-4389-b5f0-c550a2adc834 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 909.198333] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e48ece-5470-8f91-43e4-937d3673c09e, 'name': SearchDatastore_Task, 'duration_secs': 0.011806} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.199044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 909.199598] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.199671] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f72116c-76b6-4490-b040-56bbee3081be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.211816] env[65107]: DEBUG nova.compute.provider_tree [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.219674] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 909.219674] env[65107]: value = "task-5103016" [ 909.219674] env[65107]: _type = "Task" [ 909.219674] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.230103] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.265129] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.265385] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.346109] env[65107]: DEBUG nova.network.neutron [-] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 909.360665] env[65107]: DEBUG nova.network.neutron [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Successfully created port: 3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 909.461118] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.716300] env[65107]: DEBUG nova.scheduler.client.report [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 909.732624] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490235} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.733198] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.733198] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.733513] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-510eef7e-d975-4167-9f44-f2efcf350533 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.742768] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 909.742768] env[65107]: value = "task-5103017" [ 909.742768] env[65107]: _type = "Task" [ 909.742768] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.755948] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.768896] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 909.853834] env[65107]: INFO nova.compute.manager [-] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Took 1.69 seconds to deallocate network for instance. [ 910.060032] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 910.114849] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 910.115242] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 910.115900] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 910.116149] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 910.116366] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 910.116544] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 910.116869] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.117078] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 910.117257] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 910.117442] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 910.118071] env[65107]: DEBUG nova.virt.hardware [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 910.118628] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bd4725-036e-42e2-8d93-54ab1dcf744f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.130645] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbfb4b7-edb1-426a-9a04-254c564e5d8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.221765] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.195s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.224313] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.422s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.225939] env[65107]: INFO nova.compute.claims [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.247771] env[65107]: INFO nova.scheduler.client.report [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleted allocations for instance 9a46ff51-4035-4709-ae6c-0fb5397200ab [ 910.254832] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109738} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.255388] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.256216] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d4e16d-45c5-4863-a904-dbf2fbc84ee7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.280573] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.284331] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5aceacb6-ea67-47a3-82a9-5db66a20a4d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.305204] env[65107]: INFO nova.compute.manager [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Rescuing [ 910.305557] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.305730] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.305870] env[65107]: DEBUG nova.network.neutron [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 910.307987] env[65107]: DEBUG nova.compute.manager [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Received event network-changed {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 910.308168] env[65107]: DEBUG nova.compute.manager [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Refreshing instance network info cache due to event network-changed. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 910.308373] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Acquiring lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.308510] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Acquired lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.308660] env[65107]: DEBUG nova.network.neutron [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 910.320046] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 910.320046] env[65107]: value = "task-5103018" [ 910.320046] env[65107]: _type = "Task" [ 910.320046] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.325579] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.333370] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103018, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.364065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.759058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2be82be7-7ac9-42a8-b020-4b3415194486 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "9a46ff51-4035-4709-ae6c-0fb5397200ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.195s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.812161] env[65107]: WARNING neutronclient.v2_0.client [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.813105] env[65107]: WARNING openstack [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.813496] env[65107]: WARNING openstack [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.821418] env[65107]: WARNING neutronclient.v2_0.client [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.821994] env[65107]: WARNING openstack [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.822361] env[65107]: WARNING openstack [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.838839] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103018, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.887582] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.887880] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.888060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "9fa949ea-358c-46e7-a0f2-4c3275493b64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.888250] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.888426] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.893474] env[65107]: INFO nova.compute.manager [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Terminating instance [ 911.046958] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "1a813da6-f777-4888-b2ae-fbad58e01f61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.047207] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.047467] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "1a813da6-f777-4888-b2ae-fbad58e01f61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.047912] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.047912] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.055824] env[65107]: INFO nova.compute.manager [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Terminating instance [ 911.124913] env[65107]: WARNING openstack [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.125305] env[65107]: WARNING openstack [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.137162] env[65107]: DEBUG nova.network.neutron [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Successfully updated port: 3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 911.143269] env[65107]: WARNING openstack [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.143973] env[65107]: WARNING openstack [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.162649] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.164369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.164369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.164369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.164369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.165915] env[65107]: INFO nova.compute.manager [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Terminating instance [ 911.224849] env[65107]: WARNING neutronclient.v2_0.client [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.225564] env[65107]: WARNING openstack [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.226406] env[65107]: WARNING openstack [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.242828] env[65107]: WARNING neutronclient.v2_0.client [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.244139] env[65107]: WARNING openstack [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.244139] env[65107]: WARNING openstack [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.300935] env[65107]: DEBUG nova.compute.manager [req-3ab32c86-e991-42af-922a-96e55f0f7042 req-4d6dc4e0-65bd-4c11-9563-a82aeab9df92 service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Received event network-vif-plugged-3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 911.300935] env[65107]: DEBUG oslo_concurrency.lockutils [req-3ab32c86-e991-42af-922a-96e55f0f7042 req-4d6dc4e0-65bd-4c11-9563-a82aeab9df92 service nova] Acquiring lock "3a975793-f480-4f54-85c9-2c2aea11c18a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.300935] env[65107]: DEBUG oslo_concurrency.lockutils [req-3ab32c86-e991-42af-922a-96e55f0f7042 req-4d6dc4e0-65bd-4c11-9563-a82aeab9df92 service nova] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.301693] env[65107]: DEBUG oslo_concurrency.lockutils [req-3ab32c86-e991-42af-922a-96e55f0f7042 req-4d6dc4e0-65bd-4c11-9563-a82aeab9df92 service nova] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.301693] env[65107]: DEBUG nova.compute.manager [req-3ab32c86-e991-42af-922a-96e55f0f7042 req-4d6dc4e0-65bd-4c11-9563-a82aeab9df92 service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] No waiting events found dispatching network-vif-plugged-3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 911.302039] env[65107]: WARNING nova.compute.manager [req-3ab32c86-e991-42af-922a-96e55f0f7042 req-4d6dc4e0-65bd-4c11-9563-a82aeab9df92 service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Received unexpected event network-vif-plugged-3bd7ea5b-4f81-4cde-8a97-65795f27663e for instance with vm_state building and task_state spawning. [ 911.341516] env[65107]: DEBUG nova.network.neutron [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Updating instance_info_cache with network_info: [{"id": "d5026810-9268-4f3f-9e44-06ea55e41569", "address": "fa:16:3e:b7:67:65", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5026810-92", "ovs_interfaceid": "d5026810-9268-4f3f-9e44-06ea55e41569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 911.354026] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103018, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.360767] env[65107]: DEBUG nova.network.neutron [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Updating instance_info_cache with network_info: [{"id": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "address": "fa:16:3e:78:ab:3f", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.125", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03fa2dfb-b3", "ovs_interfaceid": "03fa2dfb-b370-41bb-ab18-1d3af50fd4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 911.398211] env[65107]: DEBUG nova.compute.manager [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 911.398211] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.399876] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5f2638-1347-4ded-9697-122f17ca21f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.412561] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.412922] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c06ad1d6-b534-4125-8801-281016d5c1e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.424536] env[65107]: DEBUG oslo_vmware.api [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 911.424536] env[65107]: value = "task-5103019" [ 911.424536] env[65107]: _type = "Task" [ 911.424536] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.434814] env[65107]: DEBUG oslo_vmware.api [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.471485] env[65107]: DEBUG nova.compute.manager [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Received event network-changed-190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 911.471485] env[65107]: DEBUG nova.compute.manager [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Refreshing instance network info cache due to event network-changed-190c1749-3fb8-4199-b8e7-1ab227d2b30f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 911.471485] env[65107]: DEBUG oslo_concurrency.lockutils [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Acquiring lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.471485] env[65107]: DEBUG oslo_concurrency.lockutils [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Acquired lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.471485] env[65107]: DEBUG nova.network.neutron [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Refreshing network info cache for port 190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 911.561286] env[65107]: DEBUG nova.compute.manager [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 911.561768] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.562602] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbcd65b1-8c87-4a17-a2c7-52c29cb1981b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.575052] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.575237] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e747257-3e22-4202-860f-30e6eb5194bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.585741] env[65107]: DEBUG oslo_vmware.api [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 911.585741] env[65107]: value = "task-5103020" [ 911.585741] env[65107]: _type = "Task" [ 911.585741] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.600700] env[65107]: DEBUG oslo_vmware.api [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.638717] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "refresh_cache-3a975793-f480-4f54-85c9-2c2aea11c18a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.638899] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquired lock "refresh_cache-3a975793-f480-4f54-85c9-2c2aea11c18a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.639621] env[65107]: DEBUG nova.network.neutron [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 911.672089] env[65107]: DEBUG nova.compute.manager [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 911.672364] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 911.673405] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562d60d7-63cc-4d1e-8365-7cfa9c5f7a19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.691425] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.691521] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-90296ebb-2862-447d-af96-b1b10c9b9fb0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.705109] env[65107]: DEBUG oslo_vmware.api [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 911.705109] env[65107]: value = "task-5103021" [ 911.705109] env[65107]: _type = "Task" [ 911.705109] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.715634] env[65107]: DEBUG oslo_vmware.api [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.796428] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d52da5-1a30-4333-af47-978b8edf7207 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.806141] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cb8594-cbf4-4def-b786-f235a7d544c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.846369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.853837] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3874f6f3-a774-4b7b-b189-f073774f695d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.866255] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72404ad-19d4-44bf-8115-0796dd3c7f8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.870862] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103018, 'name': ReconfigVM_Task, 'duration_secs': 1.171597} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.871456] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a06008c3-1f21-4275-b59e-bf9df2f7763c tempest-ServerExternalEventsTest-783593802 tempest-ServerExternalEventsTest-783593802-project] Releasing lock "refresh_cache-1a813da6-f777-4888-b2ae-fbad58e01f61" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.871734] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.873134] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd7db4fd-cc32-4517-a553-907406ca7db0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.888210] env[65107]: DEBUG nova.compute.provider_tree [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.895915] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 911.895915] env[65107]: value = "task-5103022" [ 911.895915] env[65107]: _type = "Task" [ 911.895915] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.908028] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103022, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.936672] env[65107]: DEBUG oslo_vmware.api [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103019, 'name': PowerOffVM_Task, 'duration_secs': 0.225426} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.936952] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.937210] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.937490] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb4ea4e5-a54a-44f1-bcc9-be5fc04fe1f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.974783] env[65107]: WARNING neutronclient.v2_0.client [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.976583] env[65107]: WARNING openstack [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.976837] env[65107]: WARNING openstack [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.009712] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.011530] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.011530] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Deleting the datastore file [datastore2] 9fa949ea-358c-46e7-a0f2-4c3275493b64 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.011530] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cc52f58-eb79-48ec-b422-6dcdb5e236ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.023978] env[65107]: DEBUG oslo_vmware.api [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 912.023978] env[65107]: value = "task-5103024" [ 912.023978] env[65107]: _type = "Task" [ 912.023978] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.046508] env[65107]: DEBUG oslo_vmware.api [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.096691] env[65107]: DEBUG oslo_vmware.api [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103020, 'name': PowerOffVM_Task, 'duration_secs': 0.211134} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.096691] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.096691] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.096691] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6543e9c4-0685-4200-8f6d-5505729450cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.127731] env[65107]: WARNING openstack [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.128706] env[65107]: WARNING openstack [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.144210] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.144704] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.217022] env[65107]: DEBUG nova.network.neutron [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 912.222455] env[65107]: DEBUG oslo_vmware.api [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103021, 'name': PowerOffVM_Task, 'duration_secs': 0.208408} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.223244] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.223575] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.223996] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5e29099-ae9d-4d15-8048-ff81d6aad3cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.245462] env[65107]: WARNING neutronclient.v2_0.client [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.247043] env[65107]: WARNING openstack [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.247043] env[65107]: WARNING openstack [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.260931] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.261383] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.335086] env[65107]: WARNING neutronclient.v2_0.client [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.335801] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.336172] env[65107]: WARNING openstack [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.374106] env[65107]: DEBUG nova.network.neutron [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updated VIF entry in instance network info cache for port 190c1749-3fb8-4199-b8e7-1ab227d2b30f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 912.374535] env[65107]: DEBUG nova.network.neutron [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updating instance_info_cache with network_info: [{"id": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "address": "fa:16:3e:39:cf:6d", "network": {"id": "6c00b9ca-3fa4-45fd-8e6b-8ee7ffebff27", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-780211679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b981c5f6b3b64476a2406cd9549f7d6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap190c1749-3f", "ovs_interfaceid": "190c1749-3fb8-4199-b8e7-1ab227d2b30f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 912.396804] env[65107]: DEBUG nova.scheduler.client.report [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 912.413098] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103022, 'name': Rename_Task, 'duration_secs': 0.172785} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.416792] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.416792] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f94072c-ef71-45dc-8e9c-6c5ae61c5984 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.425166] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 912.425166] env[65107]: value = "task-5103027" [ 912.425166] env[65107]: _type = "Task" [ 912.425166] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.439444] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103027, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.534842] env[65107]: DEBUG oslo_vmware.api [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170679} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.535319] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.535541] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 912.535719] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 912.535894] env[65107]: INFO nova.compute.manager [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Took 1.14 seconds to destroy the instance on the hypervisor. [ 912.536453] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 912.536680] env[65107]: DEBUG nova.compute.manager [-] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 912.536781] env[65107]: DEBUG nova.network.neutron [-] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 912.537058] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.537774] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.538149] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.557452] env[65107]: DEBUG nova.network.neutron [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Updating instance_info_cache with network_info: [{"id": "3bd7ea5b-4f81-4cde-8a97-65795f27663e", "address": "fa:16:3e:75:e6:d8", "network": {"id": "a262ff6b-35af-47e7-9017-9cff9d5d811d", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1160136819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05918ba13682405ba5e5e7ebde13b42b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7874ee7f-20c7-4bd8-a750-ed489e9acc65", "external-id": "nsx-vlan-transportzone-753", "segmentation_id": 753, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bd7ea5b-4f", "ovs_interfaceid": "3bd7ea5b-4f81-4cde-8a97-65795f27663e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 912.573191] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.573448] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.573630] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Deleting the datastore file [datastore1] 1a813da6-f777-4888-b2ae-fbad58e01f61 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.574048] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85649447-e4b5-4a5c-9d9f-4c5045cb3d49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.584152] env[65107]: DEBUG oslo_vmware.api [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for the task: (returnval){ [ 912.584152] env[65107]: value = "task-5103028" [ 912.584152] env[65107]: _type = "Task" [ 912.584152] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.592821] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.593138] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.593326] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleting the datastore file [datastore1] 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.594528] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20d1e642-2d8d-4c7a-9eba-66de38d21738 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.598436] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.607469] env[65107]: DEBUG oslo_vmware.api [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.614146] env[65107]: DEBUG oslo_vmware.api [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 912.614146] env[65107]: value = "task-5103029" [ 912.614146] env[65107]: _type = "Task" [ 912.614146] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.626872] env[65107]: DEBUG oslo_vmware.api [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.878671] env[65107]: DEBUG oslo_concurrency.lockutils [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] Releasing lock "refresh_cache-9fa949ea-358c-46e7-a0f2-4c3275493b64" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.880143] env[65107]: DEBUG nova.compute.manager [req-79e9fc88-1520-4343-bc16-4620ae4683e3 req-e35750c0-c179-447c-b9a4-32840228be76 service nova] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Received event network-vif-deleted-2f395da6-1303-4770-b9a5-eed2e9141181 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 912.907571] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.683s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.908172] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 912.913551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.623s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.915275] env[65107]: INFO nova.compute.claims [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.936684] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103027, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.062958] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Releasing lock "refresh_cache-3a975793-f480-4f54-85c9-2c2aea11c18a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 913.063877] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Instance network_info: |[{"id": "3bd7ea5b-4f81-4cde-8a97-65795f27663e", "address": "fa:16:3e:75:e6:d8", "network": {"id": "a262ff6b-35af-47e7-9017-9cff9d5d811d", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1160136819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05918ba13682405ba5e5e7ebde13b42b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7874ee7f-20c7-4bd8-a750-ed489e9acc65", "external-id": "nsx-vlan-transportzone-753", "segmentation_id": 753, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bd7ea5b-4f", "ovs_interfaceid": "3bd7ea5b-4f81-4cde-8a97-65795f27663e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 913.064456] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:e6:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7874ee7f-20c7-4bd8-a750-ed489e9acc65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bd7ea5b-4f81-4cde-8a97-65795f27663e', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.079091] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Creating folder: Project (05918ba13682405ba5e5e7ebde13b42b). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.079091] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f76792ff-f8ba-427d-a6ac-43ba33cf88c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.091645] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Created folder: Project (05918ba13682405ba5e5e7ebde13b42b) in parent group-v992574. [ 913.091940] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Creating folder: Instances. Parent ref: group-v992779. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.092649] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3347098a-41c7-4b16-b534-2180ef502291 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.102119] env[65107]: DEBUG oslo_vmware.api [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Task: {'id': task-5103028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274907} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.102878] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.103510] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.103756] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.104079] env[65107]: INFO nova.compute.manager [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Took 1.54 seconds to destroy the instance on the hypervisor. [ 913.104370] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 913.104663] env[65107]: DEBUG nova.compute.manager [-] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 913.104827] env[65107]: DEBUG nova.network.neutron [-] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 913.105300] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.106064] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.106356] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.117466] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Created folder: Instances in parent group-v992779. [ 913.117751] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 913.121585] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.121881] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d0bfc00-ec24-426d-b252-41624c23c810 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.142808] env[65107]: DEBUG oslo_vmware.api [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282092} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.144854] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.145084] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.145380] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.145589] env[65107]: INFO nova.compute.manager [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Took 1.47 seconds to destroy the instance on the hypervisor. [ 913.145843] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 913.146086] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.146086] env[65107]: value = "task-5103032" [ 913.146086] env[65107]: _type = "Task" [ 913.146086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.146308] env[65107]: DEBUG nova.compute.manager [-] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 913.146406] env[65107]: DEBUG nova.network.neutron [-] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 913.146718] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.147369] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.147633] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.164325] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103032, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.206259] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.245801] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.415643] env[65107]: DEBUG nova.compute.utils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 913.417874] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 913.418273] env[65107]: DEBUG nova.network.neutron [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 913.418658] env[65107]: WARNING neutronclient.v2_0.client [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.419014] env[65107]: WARNING neutronclient.v2_0.client [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.419778] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.420171] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.431856] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.432993] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-546fe999-803e-4b4c-a9be-94078d052732 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.448450] env[65107]: DEBUG oslo_vmware.api [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103027, 'name': PowerOnVM_Task, 'duration_secs': 0.71088} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.450408] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.450721] env[65107]: INFO nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Took 6.08 seconds to spawn the instance on the hypervisor. [ 913.450958] env[65107]: DEBUG nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 913.451364] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 913.451364] env[65107]: value = "task-5103033" [ 913.451364] env[65107]: _type = "Task" [ 913.451364] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.452264] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb5a75b-307e-4052-9a28-e4dd0411e115 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.457112] env[65107]: DEBUG nova.network.neutron [-] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 913.470300] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.667257] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103032, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.929310] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 913.963833] env[65107]: INFO nova.compute.manager [-] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Took 1.43 seconds to deallocate network for instance. [ 913.974574] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103033, 'name': PowerOffVM_Task, 'duration_secs': 0.295606} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.977803] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.983251] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62df748e-f706-43c7-8e29-6ddc4a9f99c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.002333] env[65107]: DEBUG nova.policy [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a974f6d9394deab176e30972bb7277', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dced4f9aef2f49cf990203b693533aa1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 914.006383] env[65107]: INFO nova.compute.manager [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Took 29.20 seconds to build instance. [ 914.036260] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e22d08-925f-4a98-a9f4-3d23c7054eac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.082646] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.083875] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-072675ab-c143-4acf-98b0-863ba7f0683f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.095646] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 914.095646] env[65107]: value = "task-5103034" [ 914.095646] env[65107]: _type = "Task" [ 914.095646] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.107962] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 914.108438] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.108779] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.109018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.109296] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.112844] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2fefe49-d6a5-4d04-aa4d-8eba0abc2e1d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.124614] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.124940] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.125925] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-743cadbb-0549-41c2-9f07-b96df80715ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.134809] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 914.134809] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c30dcf-a0ac-1ee8-7198-e2ba54ce1540" [ 914.134809] env[65107]: _type = "Task" [ 914.134809] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.144716] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c30dcf-a0ac-1ee8-7198-e2ba54ce1540, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.165827] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103032, 'name': CreateVM_Task, 'duration_secs': 0.558088} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.166092] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.166759] env[65107]: WARNING neutronclient.v2_0.client [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.167408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.167606] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.168078] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 914.171425] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-454a2f7a-ad41-4507-9a46-52409bb733a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.178211] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 914.178211] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b981c0-81ae-3bdc-01c3-e8ae96ba0614" [ 914.178211] env[65107]: _type = "Task" [ 914.178211] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.189280] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b981c0-81ae-3bdc-01c3-e8ae96ba0614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.408082] env[65107]: DEBUG nova.network.neutron [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Successfully created port: b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 914.452979] env[65107]: DEBUG nova.network.neutron [-] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.465151] env[65107]: DEBUG nova.network.neutron [-] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.473025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.479114] env[65107]: DEBUG nova.compute.manager [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Received event network-changed-3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 914.479322] env[65107]: DEBUG nova.compute.manager [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Refreshing instance network info cache due to event network-changed-3bd7ea5b-4f81-4cde-8a97-65795f27663e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 914.479498] env[65107]: DEBUG oslo_concurrency.lockutils [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Acquiring lock "refresh_cache-3a975793-f480-4f54-85c9-2c2aea11c18a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.479589] env[65107]: DEBUG oslo_concurrency.lockutils [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Acquired lock "refresh_cache-3a975793-f480-4f54-85c9-2c2aea11c18a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.479806] env[65107]: DEBUG nova.network.neutron [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Refreshing network info cache for port 3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 914.513613] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b24a2f4e-fa23-4f31-9641-b294382382ce tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "4a42392a-a4b4-43b7-a410-0b69441435be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.724s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.533835] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e92866f-c173-4901-b8ca-299ad5bb9805 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.543191] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2511b09a-31bf-4bc6-9de0-f895059c50d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.580496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dafae007-c593-44ba-a826-ed2a7a59b112 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.589423] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37f1b3a-2158-478e-9639-eb9cae2d3c86 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.605530] env[65107]: DEBUG nova.compute.provider_tree [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 914.648100] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c30dcf-a0ac-1ee8-7198-e2ba54ce1540, 'name': SearchDatastore_Task, 'duration_secs': 0.011807} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.649533] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8760933c-db94-4bfc-bd37-ede348950f49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.657490] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 914.657490] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b89fa7-6730-d893-c357-208b441b9aab" [ 914.657490] env[65107]: _type = "Task" [ 914.657490] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.668962] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b89fa7-6730-d893-c357-208b441b9aab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.691275] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b981c0-81ae-3bdc-01c3-e8ae96ba0614, 'name': SearchDatastore_Task, 'duration_secs': 0.012145} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.691514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.691754] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.692010] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.957986] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 914.960916] env[65107]: INFO nova.compute.manager [-] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Took 1.86 seconds to deallocate network for instance. [ 914.972931] env[65107]: INFO nova.compute.manager [-] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Took 1.82 seconds to deallocate network for instance. [ 914.985450] env[65107]: WARNING neutronclient.v2_0.client [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.986229] env[65107]: WARNING openstack [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.986625] env[65107]: WARNING openstack [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.997279] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 914.997718] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 914.998026] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 914.998455] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 914.998635] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 914.998855] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 914.999278] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.999602] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 914.999898] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 915.000146] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 915.000372] env[65107]: DEBUG nova.virt.hardware [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 915.001334] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900d70ef-4aec-478e-a64d-d4b65da75d1d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.012540] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da72c69-85ec-4cec-91e1-3ab12631ca55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.134303] env[65107]: ERROR nova.scheduler.client.report [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [req-1ae1369e-9ec7-475e-9987-76c8490e2395] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1ae1369e-9ec7-475e-9987-76c8490e2395"}]} [ 915.139864] env[65107]: DEBUG nova.compute.manager [req-172f615d-820f-4309-8815-a6dea171eddd req-3f170616-7e47-4993-9e3d-528a2a12df52 service nova] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Received event network-vif-deleted-190c1749-3fb8-4199-b8e7-1ab227d2b30f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 915.159364] env[65107]: DEBUG nova.scheduler.client.report [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 915.175904] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b89fa7-6730-d893-c357-208b441b9aab, 'name': SearchDatastore_Task, 'duration_secs': 0.010607} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.176165] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.176503] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. {{(pid=65107) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 915.176842] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 915.177093] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.178710] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22d45e2e-6b8f-4eca-89a6-ae64f411b375 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.181037] env[65107]: DEBUG nova.scheduler.client.report [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 915.181216] env[65107]: DEBUG nova.compute.provider_tree [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.183530] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbc0b839-e921-48c6-a0b8-bf1af66a6e2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.192676] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 915.192676] env[65107]: value = "task-5103035" [ 915.192676] env[65107]: _type = "Task" [ 915.192676] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.200714] env[65107]: DEBUG nova.scheduler.client.report [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 915.204048] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.204665] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.205706] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c6d7884-e67c-457c-bb63-69ff63cec8ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.212729] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.216475] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 915.216475] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5299d951-b1c6-349e-a0f5-6e1d3e43fc8b" [ 915.216475] env[65107]: _type = "Task" [ 915.216475] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.226524] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5299d951-b1c6-349e-a0f5-6e1d3e43fc8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.226524] env[65107]: DEBUG nova.scheduler.client.report [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 915.251829] env[65107]: WARNING openstack [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.252364] env[65107]: WARNING openstack [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.407049] env[65107]: WARNING neutronclient.v2_0.client [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.407049] env[65107]: WARNING openstack [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.407664] env[65107]: WARNING openstack [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.469694] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.477927] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.536037] env[65107]: DEBUG nova.network.neutron [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Updated VIF entry in instance network info cache for port 3bd7ea5b-4f81-4cde-8a97-65795f27663e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 915.536037] env[65107]: DEBUG nova.network.neutron [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Updating instance_info_cache with network_info: [{"id": "3bd7ea5b-4f81-4cde-8a97-65795f27663e", "address": "fa:16:3e:75:e6:d8", "network": {"id": "a262ff6b-35af-47e7-9017-9cff9d5d811d", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1160136819-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "05918ba13682405ba5e5e7ebde13b42b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7874ee7f-20c7-4bd8-a750-ed489e9acc65", "external-id": "nsx-vlan-transportzone-753", "segmentation_id": 753, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bd7ea5b-4f", "ovs_interfaceid": "3bd7ea5b-4f81-4cde-8a97-65795f27663e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.705220] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103035, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.724856] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a781c3c-fe55-4805-b32c-94f7f5fc22ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.737689] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48457bce-bd13-4c84-88c7-640d86a8d0d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.741766] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5299d951-b1c6-349e-a0f5-6e1d3e43fc8b, 'name': SearchDatastore_Task, 'duration_secs': 0.010412} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.744782] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7a611c1-3a2c-4c45-adc6-9e30ff51100e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.773803] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85423c74-e26f-404c-8f7b-b2b4189e59c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.777999] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 915.777999] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c3dcc8-fced-a356-4bd6-87cfe9312862" [ 915.777999] env[65107]: _type = "Task" [ 915.777999] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.786320] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43893f1-8161-4da5-baae-7118ee6ca172 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.796371] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c3dcc8-fced-a356-4bd6-87cfe9312862, 'name': SearchDatastore_Task, 'duration_secs': 0.012104} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.797122] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.797425] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 3a975793-f480-4f54-85c9-2c2aea11c18a/3a975793-f480-4f54-85c9-2c2aea11c18a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.797749] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29cf1311-6074-456b-b863-6d8eb8f9e2e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.809510] env[65107]: DEBUG nova.compute.provider_tree [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.818328] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 915.818328] env[65107]: value = "task-5103036" [ 915.818328] env[65107]: _type = "Task" [ 915.818328] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.830237] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.038597] env[65107]: DEBUG oslo_concurrency.lockutils [req-5534e838-bf6c-4a8d-8e54-bdd90e72a53e req-c22af1c0-c743-42f5-abfd-bb63d4dd1def service nova] Releasing lock "refresh_cache-3a975793-f480-4f54-85c9-2c2aea11c18a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 916.095775] env[65107]: DEBUG nova.network.neutron [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Successfully updated port: b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 916.209444] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.891102} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.210172] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. [ 916.211148] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e883a9-9d08-45d2-bad2-27773c4d6d3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.241903] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.243557] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34315345-68c9-4c92-abe6-37b4ca265b0b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.269496] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 916.269496] env[65107]: value = "task-5103037" [ 916.269496] env[65107]: _type = "Task" [ 916.269496] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.277869] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103037, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.334020] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103036, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.359087] env[65107]: DEBUG nova.scheduler.client.report [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 102 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 916.359380] env[65107]: DEBUG nova.compute.provider_tree [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 102 to 103 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 916.359612] env[65107]: DEBUG nova.compute.provider_tree [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 916.599241] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.599495] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.599742] env[65107]: DEBUG nova.network.neutron [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 916.698424] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52096df5-7584-38b4-953e-24ea95723e69/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 916.699477] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1017dc-92bc-41b0-bb0b-d8958f9cd46c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.714173] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52096df5-7584-38b4-953e-24ea95723e69/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 916.714390] env[65107]: ERROR oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52096df5-7584-38b4-953e-24ea95723e69/disk-0.vmdk due to incomplete transfer. [ 916.714663] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a3e4f415-d0a1-4cb8-8b58-6fb7d4ee079c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.727423] env[65107]: DEBUG oslo_vmware.rw_handles [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52096df5-7584-38b4-953e-24ea95723e69/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 916.727423] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Uploaded image 57f71a90-320c-4d88-8671-5ad259d2a173 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 916.732446] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 916.732874] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1df518f2-a6b8-4366-85e1-374af0960662 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.743146] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 916.743146] env[65107]: value = "task-5103038" [ 916.743146] env[65107]: _type = "Task" [ 916.743146] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.755106] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103038, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.777606] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103037, 'name': ReconfigVM_Task, 'duration_secs': 0.501703} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.777872] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Reconfigured VM instance instance-0000003d to attach disk [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.778821] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136a71b7-085c-4ee8-864d-36ad40f4de55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.812157] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a07dab0f-c2d3-45fa-931b-de20d842a7d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.838252] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103036, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670148} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.841779] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 3a975793-f480-4f54-85c9-2c2aea11c18a/3a975793-f480-4f54-85c9-2c2aea11c18a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.842234] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.842752] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 916.842752] env[65107]: value = "task-5103039" [ 916.842752] env[65107]: _type = "Task" [ 916.842752] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.843105] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a3499b4-f31b-40dc-9291-b8856023dec5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.858219] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.864049] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 916.864049] env[65107]: value = "task-5103040" [ 916.864049] env[65107]: _type = "Task" [ 916.864049] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.868225] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.955s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.868752] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 916.871808] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.829s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.873504] env[65107]: INFO nova.compute.claims [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 916.882818] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103040, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.885483] env[65107]: DEBUG nova.compute.manager [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Received event network-vif-deleted-03fa2dfb-b370-41bb-ab18-1d3af50fd4d9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 916.885483] env[65107]: DEBUG nova.compute.manager [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-vif-plugged-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 916.885682] env[65107]: DEBUG oslo_concurrency.lockutils [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.885884] env[65107]: DEBUG oslo_concurrency.lockutils [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.886072] env[65107]: DEBUG oslo_concurrency.lockutils [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.886292] env[65107]: DEBUG nova.compute.manager [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] No waiting events found dispatching network-vif-plugged-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 916.886560] env[65107]: WARNING nova.compute.manager [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received unexpected event network-vif-plugged-b6bba8b1-799c-4c2b-bd94-8726094948ee for instance with vm_state building and task_state spawning. [ 916.886725] env[65107]: DEBUG nova.compute.manager [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 916.886870] env[65107]: DEBUG nova.compute.manager [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing instance network info cache due to event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 916.888087] env[65107]: DEBUG oslo_concurrency.lockutils [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.995768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "d5766a03-054f-40ea-a57e-e640664ca683" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.996027] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "d5766a03-054f-40ea-a57e-e640664ca683" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.104395] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.104940] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.158196] env[65107]: DEBUG nova.network.neutron [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 917.181883] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.182272] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.254381] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103038, 'name': Destroy_Task, 'duration_secs': 0.406013} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.254746] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Destroyed the VM [ 917.254879] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 917.255159] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-351eb665-eb04-4549-af0f-28eb6f50ca90 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.263025] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 917.263025] env[65107]: value = "task-5103041" [ 917.263025] env[65107]: _type = "Task" [ 917.263025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.272080] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103041, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.273430] env[65107]: WARNING neutronclient.v2_0.client [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.274118] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.274529] env[65107]: WARNING openstack [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.309651] env[65107]: DEBUG nova.compute.manager [req-01d562eb-3b58-46c3-b3b2-f16dc46394fe req-4cbeed2b-0b9e-4cf9-9844-cbfbc2c1d822 service nova] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Received event network-vif-deleted-c9e1f243-9bcf-4a3a-8b83-1f0496d6a0b7 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 917.359340] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103039, 'name': ReconfigVM_Task, 'duration_secs': 0.22629} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.359653] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.359927] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4eb96b1a-5258-4240-afda-91a9766a854e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.370286] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 917.370286] env[65107]: value = "task-5103042" [ 917.370286] env[65107]: _type = "Task" [ 917.370286] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.380551] env[65107]: DEBUG nova.compute.utils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 917.384570] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103040, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074557} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.387879] env[65107]: DEBUG nova.network.neutron [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.392771] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 917.392997] env[65107]: DEBUG nova.network.neutron [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 917.393344] env[65107]: WARNING neutronclient.v2_0.client [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.393717] env[65107]: WARNING neutronclient.v2_0.client [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.394476] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.395000] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.402817] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.409090] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c2a36e-e203-4faa-bad2-c85f318fa7ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.416728] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103042, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.436833] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 3a975793-f480-4f54-85c9-2c2aea11c18a/3a975793-f480-4f54-85c9-2c2aea11c18a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.437681] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0facf8ce-43f3-4d3a-b666-7c954de83cff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.456123] env[65107]: DEBUG nova.policy [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 917.465484] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 917.465484] env[65107]: value = "task-5103043" [ 917.465484] env[65107]: _type = "Task" [ 917.465484] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.477066] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.500880] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 917.777398] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103041, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.816717] env[65107]: DEBUG nova.network.neutron [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Successfully created port: e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 917.882327] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103042, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.890648] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 917.905794] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.906016] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance network_info: |[{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 917.906403] env[65107]: DEBUG oslo_concurrency.lockutils [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 917.906636] env[65107]: DEBUG nova.network.neutron [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 917.908120] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:15:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6bba8b1-799c-4c2b-bd94-8726094948ee', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.917646] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating folder: Project (dced4f9aef2f49cf990203b693533aa1). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.920465] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a85e3a2d-c6d1-4ed0-b849-857bfffc5261 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.936547] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created folder: Project (dced4f9aef2f49cf990203b693533aa1) in parent group-v992574. [ 917.936776] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating folder: Instances. Parent ref: group-v992782. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.940155] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58ed1e21-8e3f-4946-8910-62ebd81042bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.952793] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created folder: Instances in parent group-v992782. [ 917.953981] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 917.953981] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.953981] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00c15606-bd48-4816-a6a3-6e4caab07ac6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.984603] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103043, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.986576] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.986576] env[65107]: value = "task-5103046" [ 917.986576] env[65107]: _type = "Task" [ 917.986576] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.993177] env[65107]: INFO nova.compute.manager [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Rebuilding instance [ 918.011299] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103046, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.036116] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 918.065069] env[65107]: DEBUG nova.compute.manager [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 918.067047] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a152f6f-678b-4ef6-b7e3-21bcb42f6aad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.277176] env[65107]: DEBUG oslo_vmware.api [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103041, 'name': RemoveSnapshot_Task, 'duration_secs': 0.902364} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.277519] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 918.277712] env[65107]: INFO nova.compute.manager [None req-e34550ca-96bc-4e4d-8a04-a27a3b8a3c50 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Took 16.07 seconds to snapshot the instance on the hypervisor. [ 918.385370] env[65107]: DEBUG oslo_vmware.api [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103042, 'name': PowerOnVM_Task, 'duration_secs': 0.516878} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.385680] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.390064] env[65107]: DEBUG nova.compute.manager [None req-98adcac7-eb2f-4cff-8c5a-ac777320bd0a tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 918.391096] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1483151d-c5d5-4d93-8dbd-f17173e7bae8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.404575] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c32534-f1dc-422a-9d7e-b7e4acccdc3f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.413659] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72228ea-9ff0-4944-b631-1fd8c358b886 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.421110] env[65107]: WARNING neutronclient.v2_0.client [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.421800] env[65107]: WARNING openstack [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.422279] env[65107]: WARNING openstack [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.466335] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe38408-9087-47aa-b712-700f58b13995 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.480122] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c71a4c-b817-40fc-9d13-55d4692a3ca5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.490793] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103043, 'name': ReconfigVM_Task, 'duration_secs': 0.778026} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.505766] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 3a975793-f480-4f54-85c9-2c2aea11c18a/3a975793-f480-4f54-85c9-2c2aea11c18a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.507075] env[65107]: DEBUG nova.compute.provider_tree [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.508786] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f694fe85-b2f2-4f94-a9fd-b69ed8e47695 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.517673] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103046, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.520515] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 918.520515] env[65107]: value = "task-5103047" [ 918.520515] env[65107]: _type = "Task" [ 918.520515] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.529576] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 918.529909] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.543291] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103047, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.629117] env[65107]: WARNING openstack [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.629527] env[65107]: WARNING openstack [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.718568] env[65107]: WARNING neutronclient.v2_0.client [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.719368] env[65107]: WARNING openstack [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.719734] env[65107]: WARNING openstack [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.812314] env[65107]: DEBUG nova.network.neutron [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updated VIF entry in instance network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 918.812699] env[65107]: DEBUG nova.network.neutron [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 918.905785] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 918.934370] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 918.934651] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 918.934807] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 918.934987] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 918.935143] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 918.935288] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 918.935794] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.935953] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 918.936148] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 918.936315] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 918.936490] env[65107]: DEBUG nova.virt.hardware [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 918.937700] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c9e0d5-8926-4fd6-85d4-6e83a413197a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.947044] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab0f401-0d63-427f-bf87-c3a9c71486d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.004050] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103046, 'name': CreateVM_Task, 'duration_secs': 0.602374} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.004050] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.004652] env[65107]: WARNING neutronclient.v2_0.client [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.005126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.005379] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.005851] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 919.006236] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89a84e60-86f3-4291-af17-d2cf1433d2d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.017270] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 919.017270] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8539-2094-88c4-25fa-5fbcd49ce738" [ 919.017270] env[65107]: _type = "Task" [ 919.017270] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.030010] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed8539-2094-88c4-25fa-5fbcd49ce738, 'name': SearchDatastore_Task, 'duration_secs': 0.013131} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.030800] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.031069] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.031575] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.031575] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.031698] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.032862] env[65107]: ERROR nova.scheduler.client.report [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [req-fb6c7cf3-539c-4f26-aa7d-7639dff23904] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fb6c7cf3-539c-4f26-aa7d-7639dff23904"}]} [ 919.033371] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5aae9c4c-a01e-481a-bc70-69c84194a44e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.041959] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 919.045145] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103047, 'name': Rename_Task, 'duration_secs': 0.154282} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.046456] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.047117] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-012bde73-9131-4877-a45f-cbdde35ab8ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.054908] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.055117] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.056977] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c6a4a18-2318-4775-9292-c446385c9fd5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.059493] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 919.059493] env[65107]: value = "task-5103048" [ 919.059493] env[65107]: _type = "Task" [ 919.059493] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.066589] env[65107]: DEBUG nova.scheduler.client.report [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 919.069728] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 919.069728] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a328d-8254-5388-1ab9-732eeee050cb" [ 919.069728] env[65107]: _type = "Task" [ 919.069728] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.077450] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103048, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.084413] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a328d-8254-5388-1ab9-732eeee050cb, 'name': SearchDatastore_Task, 'duration_secs': 0.011559} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.086698] env[65107]: DEBUG nova.scheduler.client.report [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 919.086698] env[65107]: DEBUG nova.compute.provider_tree [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 919.088873] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e46da8ef-73b1-4e06-a5c6-b1c7a50120bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.094511] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.094946] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5a82ea5-cd99-4ae0-94ae-bf74c902a7e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.102412] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 919.102412] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2443f-15ec-c689-47f2-eab141ed2972" [ 919.102412] env[65107]: _type = "Task" [ 919.102412] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.107847] env[65107]: DEBUG nova.scheduler.client.report [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 919.111576] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 919.111576] env[65107]: value = "task-5103049" [ 919.111576] env[65107]: _type = "Task" [ 919.111576] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.121117] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2443f-15ec-c689-47f2-eab141ed2972, 'name': SearchDatastore_Task, 'duration_secs': 0.011776} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.121913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.122259] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.122563] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca999655-1229-4466-9849-2c8f1bd68296 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.129040] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103049, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.135288] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 919.135288] env[65107]: value = "task-5103050" [ 919.135288] env[65107]: _type = "Task" [ 919.135288] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.140684] env[65107]: DEBUG nova.scheduler.client.report [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 919.147256] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103050, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.315506] env[65107]: DEBUG oslo_concurrency.lockutils [req-4bdc0a9d-87d1-4bbd-9c05-95c79c9a5d72 req-40174a6b-9328-42c7-b25e-f9810b418d5d service nova] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.386859] env[65107]: DEBUG nova.compute.manager [req-40db4b71-d260-4a1d-bdf1-5bd0d43d541b req-593bb493-7ed6-42b9-8e92-acb4845bfe8d service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-vif-plugged-e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 919.387115] env[65107]: DEBUG oslo_concurrency.lockutils [req-40db4b71-d260-4a1d-bdf1-5bd0d43d541b req-593bb493-7ed6-42b9-8e92-acb4845bfe8d service nova] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.387327] env[65107]: DEBUG oslo_concurrency.lockutils [req-40db4b71-d260-4a1d-bdf1-5bd0d43d541b req-593bb493-7ed6-42b9-8e92-acb4845bfe8d service nova] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.387495] env[65107]: DEBUG oslo_concurrency.lockutils [req-40db4b71-d260-4a1d-bdf1-5bd0d43d541b req-593bb493-7ed6-42b9-8e92-acb4845bfe8d service nova] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.387658] env[65107]: DEBUG nova.compute.manager [req-40db4b71-d260-4a1d-bdf1-5bd0d43d541b req-593bb493-7ed6-42b9-8e92-acb4845bfe8d service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] No waiting events found dispatching network-vif-plugged-e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 919.387820] env[65107]: WARNING nova.compute.manager [req-40db4b71-d260-4a1d-bdf1-5bd0d43d541b req-593bb493-7ed6-42b9-8e92-acb4845bfe8d service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received unexpected event network-vif-plugged-e5280424-9c40-4f31-a39e-260899363c1d for instance with vm_state building and task_state spawning. [ 919.507660] env[65107]: DEBUG nova.network.neutron [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Successfully updated port: e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 919.567143] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.572426] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103048, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.624438] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103049, 'name': PowerOffVM_Task, 'duration_secs': 0.183387} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.625746] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.626448] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.627316] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d56231-6af1-4ec5-8558-0a28074923c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.630586] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39bd4ea-9801-43f0-97bb-0f6a446ec3f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.644718] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.645543] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1ab5f89-8d8c-4bd0-9f44-ded0087b4c57 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.648800] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee2cd97-0142-44e9-8102-52434b00f3b4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.655502] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103050, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.690586] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1411c57-a232-4d35-b920-edaa8ca55a24 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.693917] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.694290] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.694593] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Deleting the datastore file [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.695027] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab740e7c-380b-4324-be68-9b14b6a61903 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.708412] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9e7cdc-e004-4f3c-833b-90712edcb74c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.713627] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 919.713627] env[65107]: value = "task-5103052" [ 919.713627] env[65107]: _type = "Task" [ 919.713627] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.725707] env[65107]: DEBUG nova.compute.provider_tree [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 919.734694] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103052, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.013475] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.013908] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.014127] env[65107]: DEBUG nova.network.neutron [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 920.074756] env[65107]: DEBUG oslo_vmware.api [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103048, 'name': PowerOnVM_Task, 'duration_secs': 0.859545} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.075094] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.075319] env[65107]: INFO nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Took 10.02 seconds to spawn the instance on the hypervisor. [ 920.075503] env[65107]: DEBUG nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 920.076568] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65755514-fcbd-4eb9-86e1-1194e259dd93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.148607] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103050, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598669} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.148920] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.149162] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.149515] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1bf4e131-1cab-45a5-b5b4-9ee827b998a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.158330] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 920.158330] env[65107]: value = "task-5103053" [ 920.158330] env[65107]: _type = "Task" [ 920.158330] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.169771] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103053, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.234383] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103052, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180793} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.234509] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.234815] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.235094] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.266504] env[65107]: DEBUG nova.scheduler.client.report [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 920.266783] env[65107]: DEBUG nova.compute.provider_tree [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 105 to 106 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 920.266969] env[65107]: DEBUG nova.compute.provider_tree [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 920.274024] env[65107]: INFO nova.compute.manager [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Unrescuing [ 920.274024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.274024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.274024] env[65107]: DEBUG nova.network.neutron [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 920.518226] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.519487] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.557479] env[65107]: DEBUG nova.network.neutron [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 920.578584] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.578976] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.595355] env[65107]: INFO nova.compute.manager [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Took 30.03 seconds to build instance. [ 920.643579] env[65107]: WARNING neutronclient.v2_0.client [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.644283] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.644639] env[65107]: WARNING openstack [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.670184] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103053, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088993} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.670496] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.671425] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253f8db4-f317-4273-a90f-c9614816aae7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.696856] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.701790] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6adcf099-a9be-46a9-ae2b-9764b481b683 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.727589] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 920.727589] env[65107]: value = "task-5103054" [ 920.727589] env[65107]: _type = "Task" [ 920.727589] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.737576] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103054, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.756808] env[65107]: DEBUG nova.network.neutron [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 920.773102] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.901s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.774161] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 920.778028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.291s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.779715] env[65107]: INFO nova.compute.claims [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.782970] env[65107]: WARNING neutronclient.v2_0.client [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.783710] env[65107]: WARNING openstack [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.784107] env[65107]: WARNING openstack [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.913580] env[65107]: WARNING openstack [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.913996] env[65107]: WARNING openstack [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.010271] env[65107]: WARNING neutronclient.v2_0.client [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.010956] env[65107]: WARNING openstack [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.011396] env[65107]: WARNING openstack [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.098714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afe8c808-8b87-44e5-ac7d-4d56a102fd96 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.539s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.128078] env[65107]: DEBUG nova.network.neutron [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Updating instance_info_cache with network_info: [{"id": "d5026810-9268-4f3f-9e44-06ea55e41569", "address": "fa:16:3e:b7:67:65", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5026810-92", "ovs_interfaceid": "d5026810-9268-4f3f-9e44-06ea55e41569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 921.238023] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103054, 'name': ReconfigVM_Task, 'duration_secs': 0.31159} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.238141] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfigured VM instance instance-00000048 to attach disk [datastore2] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.238741] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7777c3a0-74fe-4c0b-8dec-414dc7c9a1fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.247680] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 921.247680] env[65107]: value = "task-5103055" [ 921.247680] env[65107]: _type = "Task" [ 921.247680] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.256702] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103055, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.259321] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.259664] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Instance network_info: |[{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 921.260119] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:63:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3e0aae3-33d1-403b-bfaf-306f77a1422e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5280424-9c40-4f31-a39e-260899363c1d', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.268051] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 921.270338] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.271074] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bcabcf0-8329-4658-acf0-7f47a8f981ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.289308] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 921.289604] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 921.289700] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 921.289875] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 921.290030] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 921.290181] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 921.290412] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.290594] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 921.290764] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 921.290922] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 921.291108] env[65107]: DEBUG nova.virt.hardware [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 921.292645] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50de1bdf-ff2c-4e1d-ac27-330d163f7070 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.296147] env[65107]: DEBUG nova.compute.utils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 921.299503] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 921.299724] env[65107]: DEBUG nova.network.neutron [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 921.300034] env[65107]: WARNING neutronclient.v2_0.client [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.300390] env[65107]: WARNING neutronclient.v2_0.client [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.300970] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.301337] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.312151] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.312151] env[65107]: value = "task-5103056" [ 921.312151] env[65107]: _type = "Task" [ 921.312151] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.320968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46745a4e-3a5b-4ab4-a71c-d6d417b48d3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.330395] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103056, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.341975] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.347940] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 921.348726] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.348968] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f102497-84bc-49e8-981b-59e37f67a77b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.368038] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.368038] env[65107]: value = "task-5103057" [ 921.368038] env[65107]: _type = "Task" [ 921.368038] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.377377] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103057, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.381033] env[65107]: DEBUG nova.policy [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '365eed31877241e58fc860ed5620d5f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c805fe68d0fd4147b9b4b7fdc9d6a6d0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 921.430176] env[65107]: DEBUG nova.compute.manager [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-changed-e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 921.430602] env[65107]: DEBUG nova.compute.manager [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing instance network info cache due to event network-changed-e5280424-9c40-4f31-a39e-260899363c1d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 921.430687] env[65107]: DEBUG oslo_concurrency.lockutils [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.430836] env[65107]: DEBUG oslo_concurrency.lockutils [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.430992] env[65107]: DEBUG nova.network.neutron [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing network info cache for port e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 921.631532] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-ee50d08a-57fc-4c05-96fe-a11fe2708165" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.634031] env[65107]: DEBUG nova.objects.instance [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'flavor' on Instance uuid ee50d08a-57fc-4c05-96fe-a11fe2708165 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.694650] env[65107]: DEBUG nova.network.neutron [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Successfully created port: b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 921.759309] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103055, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.812717] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 921.828877] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103056, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.881791] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103057, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.935036] env[65107]: WARNING neutronclient.v2_0.client [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.935036] env[65107]: WARNING openstack [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.935274] env[65107]: WARNING openstack [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.010779] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "3a975793-f480-4f54-85c9-2c2aea11c18a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.010923] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.011131] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "3a975793-f480-4f54-85c9-2c2aea11c18a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.011342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.011526] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.022263] env[65107]: INFO nova.compute.manager [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Terminating instance [ 922.057451] env[65107]: WARNING openstack [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.058058] env[65107]: WARNING openstack [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.144197] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba710b1a-3dcc-4a2f-bfba-4ea127dd7be3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.148659] env[65107]: WARNING neutronclient.v2_0.client [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.149396] env[65107]: WARNING openstack [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.149889] env[65107]: WARNING openstack [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.180268] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.183668] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6118d751-02e6-4daa-81b0-4d51fa8b2a4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.191638] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 922.191638] env[65107]: value = "task-5103058" [ 922.191638] env[65107]: _type = "Task" [ 922.191638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.201289] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103058, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.264922] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103055, 'name': Rename_Task, 'duration_secs': 0.892674} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.270110] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.278869] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5a2b3df-288d-43b3-88c2-aa2a3304db8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.290332] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 922.290332] env[65107]: value = "task-5103059" [ 922.290332] env[65107]: _type = "Task" [ 922.290332] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.303713] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103059, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.335262] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103056, 'name': CreateVM_Task, 'duration_secs': 0.733806} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.338230] env[65107]: DEBUG nova.network.neutron [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updated VIF entry in instance network info cache for port e5280424-9c40-4f31-a39e-260899363c1d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 922.338893] env[65107]: DEBUG nova.network.neutron [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 922.340356] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.344325] env[65107]: WARNING neutronclient.v2_0.client [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.344325] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.344325] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.344325] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 922.345055] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79c3bdb0-0779-4436-afb9-33c862b5a687 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.351449] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 922.351449] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52955808-c93c-5971-fede-b68b9771f754" [ 922.351449] env[65107]: _type = "Task" [ 922.351449] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.359170] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dd2f3a-ab4e-4b79-8b62-93b4bcb58518 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.366320] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52955808-c93c-5971-fede-b68b9771f754, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.375506] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e01f27f-170d-44a5-9781-d5572d94b31f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.385779] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103057, 'name': CreateVM_Task, 'duration_secs': 0.893221} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.414051] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.415291] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.416754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62815d68-c98b-4e4f-9a9a-20a66abe6a6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.428807] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b949dc-2977-4b90-ab03-53e50d42f7ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.447091] env[65107]: DEBUG nova.compute.provider_tree [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.526935] env[65107]: DEBUG nova.compute.manager [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 922.527249] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.528237] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab751662-3fcb-47bd-8914-070f9942278e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.537379] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.537677] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35e7043a-45f5-441b-b13c-e7a67b35bcb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.545555] env[65107]: DEBUG oslo_vmware.api [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 922.545555] env[65107]: value = "task-5103060" [ 922.545555] env[65107]: _type = "Task" [ 922.545555] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.554900] env[65107]: DEBUG oslo_vmware.api [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.702226] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103058, 'name': PowerOffVM_Task, 'duration_secs': 0.248833} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.702530] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.708880] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 922.709272] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7f513db-f923-4583-830e-1e7934de73a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.734804] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 922.734804] env[65107]: value = "task-5103061" [ 922.734804] env[65107]: _type = "Task" [ 922.734804] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.747093] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103061, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.802525] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103059, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.831340] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 922.846524] env[65107]: DEBUG oslo_concurrency.lockutils [req-eedfee0c-9e8d-4ec6-8fef-c1cc57045259 req-fdf0363d-324e-4de6-85b4-99d7954345f0 service nova] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.864175] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52955808-c93c-5971-fede-b68b9771f754, 'name': SearchDatastore_Task, 'duration_secs': 0.013084} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.866490] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 922.866740] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 922.866895] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 922.867088] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 922.867233] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 922.867378] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 922.867587] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.867789] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 922.867964] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 922.868140] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 922.868316] env[65107]: DEBUG nova.virt.hardware [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 922.868655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.868867] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.869111] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.869254] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.869426] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.871427] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef42abbe-e05e-4ee2-bd91-201121d3b0ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.874124] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.874448] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 922.874721] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ca36e6b-4eef-484c-a2ec-f36eef88033c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.877622] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c96874b4-74e9-4dc3-a952-70a65ee1e53d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.887572] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef6cbc3-5c11-49b9-bdd5-311fa8fa0e97 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.892725] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 922.892725] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527fa7af-af6d-5ac8-db52-1aca401ff02c" [ 922.892725] env[65107]: _type = "Task" [ 922.892725] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.894699] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.894927] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.900093] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5579a7d-5d53-413b-9c24-ebf9b3a7a8db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.919953] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 922.919953] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5255979a-7df9-9ebd-5135-6f3f56f4f416" [ 922.919953] env[65107]: _type = "Task" [ 922.919953] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.920202] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527fa7af-af6d-5ac8-db52-1aca401ff02c, 'name': SearchDatastore_Task, 'duration_secs': 0.010182} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.920583] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.920818] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.921038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.930347] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5255979a-7df9-9ebd-5135-6f3f56f4f416, 'name': SearchDatastore_Task} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.931410] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e0fe738-ed4b-4845-ad7e-122fa20b3714 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.937524] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 922.937524] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bae473-0252-ac18-ade7-066a5a1a0081" [ 922.937524] env[65107]: _type = "Task" [ 922.937524] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.948103] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bae473-0252-ac18-ade7-066a5a1a0081, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.985542] env[65107]: DEBUG nova.scheduler.client.report [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 106 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 922.985961] env[65107]: DEBUG nova.compute.provider_tree [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 106 to 107 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 922.986265] env[65107]: DEBUG nova.compute.provider_tree [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.057222] env[65107]: DEBUG oslo_vmware.api [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103060, 'name': PowerOffVM_Task, 'duration_secs': 0.231306} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.058026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 923.058026] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 923.058026] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30b5c204-a2ca-4fd0-b576-588fbe2abd8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.131184] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 923.133234] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 923.133234] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Deleting the datastore file [datastore2] 3a975793-f480-4f54-85c9-2c2aea11c18a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.133234] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02f86a79-588b-4df0-b5da-7ee64bf7405b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.141943] env[65107]: DEBUG oslo_vmware.api [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for the task: (returnval){ [ 923.141943] env[65107]: value = "task-5103063" [ 923.141943] env[65107]: _type = "Task" [ 923.141943] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.151142] env[65107]: DEBUG oslo_vmware.api [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.195163] env[65107]: DEBUG nova.compute.manager [req-bee269ac-ea6d-4149-9f60-16d4b064509a req-7ef0c4ec-8043-467b-b4d3-3213ab572b77 service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Received event network-vif-plugged-b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 923.195453] env[65107]: DEBUG oslo_concurrency.lockutils [req-bee269ac-ea6d-4149-9f60-16d4b064509a req-7ef0c4ec-8043-467b-b4d3-3213ab572b77 service nova] Acquiring lock "fead9fe9-23f4-4d13-bf44-be0727057ddd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 923.195679] env[65107]: DEBUG oslo_concurrency.lockutils [req-bee269ac-ea6d-4149-9f60-16d4b064509a req-7ef0c4ec-8043-467b-b4d3-3213ab572b77 service nova] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 923.195944] env[65107]: DEBUG oslo_concurrency.lockutils [req-bee269ac-ea6d-4149-9f60-16d4b064509a req-7ef0c4ec-8043-467b-b4d3-3213ab572b77 service nova] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.196187] env[65107]: DEBUG nova.compute.manager [req-bee269ac-ea6d-4149-9f60-16d4b064509a req-7ef0c4ec-8043-467b-b4d3-3213ab572b77 service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] No waiting events found dispatching network-vif-plugged-b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 923.196398] env[65107]: WARNING nova.compute.manager [req-bee269ac-ea6d-4149-9f60-16d4b064509a req-7ef0c4ec-8043-467b-b4d3-3213ab572b77 service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Received unexpected event network-vif-plugged-b58f1ae3-f99e-46a9-81be-e3d4925f3f85 for instance with vm_state building and task_state spawning. [ 923.249944] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103061, 'name': ReconfigVM_Task, 'duration_secs': 0.253644} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.250264] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 923.250449] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.250764] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91ee3fcb-bb6f-43cd-8683-3f9352c1bd4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.259020] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 923.259020] env[65107]: value = "task-5103064" [ 923.259020] env[65107]: _type = "Task" [ 923.259020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.267492] env[65107]: DEBUG nova.network.neutron [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Successfully updated port: b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 923.274299] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103064, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.302455] env[65107]: DEBUG oslo_vmware.api [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103059, 'name': PowerOnVM_Task, 'duration_secs': 0.517277} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.302797] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.303014] env[65107]: INFO nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Took 8.34 seconds to spawn the instance on the hypervisor. [ 923.303208] env[65107]: DEBUG nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 923.304228] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1e02b4-217c-4129-a6f4-1796d5fd9d10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.449052] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bae473-0252-ac18-ade7-066a5a1a0081, 'name': SearchDatastore_Task, 'duration_secs': 0.010767} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.449052] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 923.449276] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] afa1a4cc-aa43-4e10-982f-edaf5cdee584/afa1a4cc-aa43-4e10-982f-edaf5cdee584.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.449530] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.449604] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.449805] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2dca3701-a7b5-482b-9da1-1dc0cc9cf36d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.451884] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe15d61c-6d5b-43d7-9cb9-7c335555d593 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.459358] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 923.459358] env[65107]: value = "task-5103065" [ 923.459358] env[65107]: _type = "Task" [ 923.459358] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.464234] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.465499] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.465687] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-832b2b21-c638-4f4a-a512-da8c0ac23f13 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.471052] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.474513] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 923.474513] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527abc16-b729-4250-9930-fb275b5d2d0c" [ 923.474513] env[65107]: _type = "Task" [ 923.474513] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.483651] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527abc16-b729-4250-9930-fb275b5d2d0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.492198] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.714s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.492792] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 923.495634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.694s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 923.653068] env[65107]: DEBUG oslo_vmware.api [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Task: {'id': task-5103063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143634} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.653068] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.653068] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.653637] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.653637] env[65107]: INFO nova.compute.manager [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 923.653851] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 923.654106] env[65107]: DEBUG nova.compute.manager [-] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 923.654239] env[65107]: DEBUG nova.network.neutron [-] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 923.654558] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.655187] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.655504] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.702293] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.770838] env[65107]: DEBUG oslo_vmware.api [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103064, 'name': PowerOnVM_Task, 'duration_secs': 0.385594} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.771158] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.771498] env[65107]: DEBUG nova.compute.manager [None req-a72aeb77-015b-4247-8fdc-ff5f491d0ed9 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 923.772474] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b84046-5ae2-442f-b7f6-6d1419d131c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.776485] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-fead9fe9-23f4-4d13-bf44-be0727057ddd" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.776590] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-fead9fe9-23f4-4d13-bf44-be0727057ddd" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.776843] env[65107]: DEBUG nova.network.neutron [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 923.825573] env[65107]: INFO nova.compute.manager [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Took 28.04 seconds to build instance. [ 923.971231] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103065, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.986445] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527abc16-b729-4250-9930-fb275b5d2d0c, 'name': SearchDatastore_Task, 'duration_secs': 0.009654} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.987333] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fa8fb9f-78e0-4e23-9d09-b02ecd3d8f98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.993584] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 923.993584] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2c2ee-214b-fa53-0c3e-7f7d26d0c5dd" [ 923.993584] env[65107]: _type = "Task" [ 923.993584] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.015018] env[65107]: DEBUG nova.compute.utils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 924.019344] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2c2ee-214b-fa53-0c3e-7f7d26d0c5dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.019344] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 924.019344] env[65107]: DEBUG nova.network.neutron [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 924.019344] env[65107]: WARNING neutronclient.v2_0.client [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.020162] env[65107]: WARNING neutronclient.v2_0.client [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.020750] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.021366] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.149038] env[65107]: DEBUG nova.policy [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d657d580aff4166922a56e46f900135', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75f92ee058014228898c5050f167f31c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 924.283532] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.283532] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.328610] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d97c99d2-17f2-479f-90eb-b914abbd55e6 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.554s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.457239] env[65107]: DEBUG nova.network.neutron [-] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.461296] env[65107]: DEBUG nova.network.neutron [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Successfully created port: 205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 924.481787] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103065, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521452} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.482839] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] afa1a4cc-aa43-4e10-982f-edaf5cdee584/afa1a4cc-aa43-4e10-982f-edaf5cdee584.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.483070] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.483359] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf296969-9066-44b0-b587-07311e085427 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.492565] env[65107]: DEBUG nova.network.neutron [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 924.503366] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 924.503366] env[65107]: value = "task-5103066" [ 924.503366] env[65107]: _type = "Task" [ 924.503366] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.512440] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2c2ee-214b-fa53-0c3e-7f7d26d0c5dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010125} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.513223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.513384] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.513764] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-566ff8f1-d830-49dc-bd76-bbc9c5c395a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.519378] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 924.524049] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Applying migration context for instance 71e6e280-50c3-4655-a6a3-40ea4c650d96 as it has an incoming, in-progress migration cb52d55b-9ce2-4ccf-996d-1ceb56e3a62b. Migration status is confirming {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 924.527882] env[65107]: INFO nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating resource usage from migration cb52d55b-9ce2-4ccf-996d-1ceb56e3a62b [ 924.528993] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103066, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.530715] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.531122] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.545058] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 924.545058] env[65107]: value = "task-5103067" [ 924.545058] env[65107]: _type = "Task" [ 924.545058] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.553950] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.556319] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.556492] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bfe544ec-9b08-4118-a940-a51520ecaac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.556626] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 82e8c093-b718-4d38-9682-ba2e710e4b93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.556743] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.556856] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 179e613b-e5a2-4fbc-8fa6-b72769425ff3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.556977] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 40474981-eeef-492e-8a8b-aaea7f554c02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.557128] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.557266] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 88cac920-3c1e-41dc-9e52-f51f48a80f56 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.557454] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ee50d08a-57fc-4c05-96fe-a11fe2708165 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.557655] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.557842] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 40e17da0-b459-4619-8e9e-18c3093328c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.558080] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3f34d6a2-0370-4239-9556-b4333e454d58 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.558386] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 9fa949ea-358c-46e7-a0f2-4c3275493b64 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.558513] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance fb4229c6-b80d-49e9-a68b-baa12c0004ed is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.558609] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Migration cb52d55b-9ce2-4ccf-996d-1ceb56e3a62b is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 924.558749] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 71e6e280-50c3-4655-a6a3-40ea4c650d96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.558842] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cddb14db-34c3-4783-b21e-e0b4623a3a9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.558952] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ac631507-8abf-4019-bdd5-7684dd51d1e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.559087] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 1a813da6-f777-4888-b2ae-fbad58e01f61 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 924.559207] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 4a42392a-a4b4-43b7-a410-0b69441435be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.559709] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3a975793-f480-4f54-85c9-2c2aea11c18a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.559890] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d2e7eee7-c9b2-49aa-92f9-1efd7c901509 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.560026] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance afa1a4cc-aa43-4e10-982f-edaf5cdee584 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.560143] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance fead9fe9-23f4-4d13-bf44-be0727057ddd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.560256] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 70172f4f-2658-4bc1-857f-0dffde2d5ab0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 924.677918] env[65107]: WARNING neutronclient.v2_0.client [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.678635] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.678982] env[65107]: WARNING openstack [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.787357] env[65107]: DEBUG nova.network.neutron [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Updating instance_info_cache with network_info: [{"id": "b58f1ae3-f99e-46a9-81be-e3d4925f3f85", "address": "fa:16:3e:b0:83:30", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb58f1ae3-f9", "ovs_interfaceid": "b58f1ae3-f99e-46a9-81be-e3d4925f3f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.843617] env[65107]: DEBUG nova.compute.manager [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 924.843617] env[65107]: DEBUG nova.compute.manager [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing instance network info cache due to event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 924.843822] env[65107]: DEBUG oslo_concurrency.lockutils [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.846928] env[65107]: DEBUG oslo_concurrency.lockutils [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.847184] env[65107]: DEBUG nova.network.neutron [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 924.964021] env[65107]: INFO nova.compute.manager [-] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Took 1.31 seconds to deallocate network for instance. [ 925.015972] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103066, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083715} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.016578] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.017466] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192a3494-01c7-4317-ab4c-298a0734345d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.051502] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] afa1a4cc-aa43-4e10-982f-edaf5cdee584/afa1a4cc-aa43-4e10-982f-edaf5cdee584.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.052256] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2329898-2819-4873-a9ce-04e9fed86e92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.076695] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 888d60d4-bc9c-48af-97f9-434883884e8a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 925.087885] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103067, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.090708] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 925.090708] env[65107]: value = "task-5103068" [ 925.090708] env[65107]: _type = "Task" [ 925.090708] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.103245] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103068, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.296110] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-fead9fe9-23f4-4d13-bf44-be0727057ddd" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.296253] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance network_info: |[{"id": "b58f1ae3-f99e-46a9-81be-e3d4925f3f85", "address": "fa:16:3e:b0:83:30", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb58f1ae3-f9", "ovs_interfaceid": "b58f1ae3-f99e-46a9-81be-e3d4925f3f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 925.296758] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:83:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b58f1ae3-f99e-46a9-81be-e3d4925f3f85', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.305439] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating folder: Project (c805fe68d0fd4147b9b4b7fdc9d6a6d0). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.306210] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7cb1345-a8c0-441a-8a8e-5a3d1a7c3f11 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.317766] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created folder: Project (c805fe68d0fd4147b9b4b7fdc9d6a6d0) in parent group-v992574. [ 925.317991] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating folder: Instances. Parent ref: group-v992787. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.318282] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52c70264-adde-40a2-8bed-cb52ceef3bbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.330918] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created folder: Instances in parent group-v992787. [ 925.331262] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 925.332035] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.332035] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8c88c8b-bb77-4ab5-99a2-eda6850a8aa1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.350592] env[65107]: WARNING neutronclient.v2_0.client [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.351283] env[65107]: WARNING openstack [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.351645] env[65107]: WARNING openstack [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.362586] env[65107]: DEBUG nova.compute.manager [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Received event network-changed-b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.362586] env[65107]: DEBUG nova.compute.manager [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Refreshing instance network info cache due to event network-changed-b58f1ae3-f99e-46a9-81be-e3d4925f3f85. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 925.362586] env[65107]: DEBUG oslo_concurrency.lockutils [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Acquiring lock "refresh_cache-fead9fe9-23f4-4d13-bf44-be0727057ddd" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.362735] env[65107]: DEBUG oslo_concurrency.lockutils [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Acquired lock "refresh_cache-fead9fe9-23f4-4d13-bf44-be0727057ddd" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.362812] env[65107]: DEBUG nova.network.neutron [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Refreshing network info cache for port b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 925.369025] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.369025] env[65107]: value = "task-5103071" [ 925.369025] env[65107]: _type = "Task" [ 925.369025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.377689] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103071, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.470693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.502522] env[65107]: WARNING openstack [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.503282] env[65107]: WARNING openstack [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.554220] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 925.567427] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103067, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.753576} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.567735] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.567948] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.568285] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18434065-e36f-480d-8089-78b1fe56bffd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.579295] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 925.579295] env[65107]: value = "task-5103072" [ 925.579295] env[65107]: _type = "Task" [ 925.579295] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.585702] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 925.600727] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 925.601176] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 925.601407] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 925.601598] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 925.601745] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 925.601889] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 925.602115] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.602275] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 925.602701] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 925.602935] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 925.603214] env[65107]: DEBUG nova.virt.hardware [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 925.608053] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebad4793-6971-479d-852d-89e8a446211d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.612057] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.614291] env[65107]: WARNING neutronclient.v2_0.client [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.614938] env[65107]: WARNING openstack [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.615297] env[65107]: WARNING openstack [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.630505] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103068, 'name': ReconfigVM_Task, 'duration_secs': 0.5056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.632888] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfigured VM instance instance-00000049 to attach disk [datastore1] afa1a4cc-aa43-4e10-982f-edaf5cdee584/afa1a4cc-aa43-4e10-982f-edaf5cdee584.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.633684] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aac2fc60-f46b-4aa1-8e64-10143b9ad79f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.636563] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6620e77-3767-4327-80d0-51bf0c5dc4e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.655589] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 925.655589] env[65107]: value = "task-5103073" [ 925.655589] env[65107]: _type = "Task" [ 925.655589] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.666467] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103073, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.721123] env[65107]: DEBUG nova.network.neutron [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updated VIF entry in instance network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 925.721123] env[65107]: DEBUG nova.network.neutron [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 925.867415] env[65107]: WARNING neutronclient.v2_0.client [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.868503] env[65107]: WARNING openstack [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.869102] env[65107]: WARNING openstack [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.895097] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103071, 'name': CreateVM_Task, 'duration_secs': 0.433597} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.895396] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.896144] env[65107]: WARNING neutronclient.v2_0.client [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.896668] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.897093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.897431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 925.898289] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cef70fa-f24f-46c3-8949-bd62a73ec2c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.905527] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 925.905527] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b922f8-b2a6-f08d-e77a-8eced7794ccf" [ 925.905527] env[65107]: _type = "Task" [ 925.905527] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.922280] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b922f8-b2a6-f08d-e77a-8eced7794ccf, 'name': SearchDatastore_Task, 'duration_secs': 0.010882} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.922280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.922280] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.922280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.922280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.922280] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.922280] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ec743ba-a0ce-4c6e-ac19-fdddc8449276 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.930578] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.930759] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.931559] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b9d65ab-9071-4d05-8367-57d0e63c996f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.940172] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 925.940172] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5278f532-b231-7826-a28e-9e4e789afdcb" [ 925.940172] env[65107]: _type = "Task" [ 925.940172] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.953025] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5278f532-b231-7826-a28e-9e4e789afdcb, 'name': SearchDatastore_Task, 'duration_secs': 0.010555} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.954163] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7005007c-69af-45b7-b3de-788046ae36d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.960671] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 925.960671] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b27c2b-0863-b31f-adcc-3c5e843f8fc9" [ 925.960671] env[65107]: _type = "Task" [ 925.960671] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.969294] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b27c2b-0863-b31f-adcc-3c5e843f8fc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.024907] env[65107]: WARNING openstack [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.025332] env[65107]: WARNING openstack [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.053981] env[65107]: DEBUG nova.network.neutron [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Successfully updated port: 205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 926.092607] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d5766a03-054f-40ea-a57e-e640664ca683 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 926.093483] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145629} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.093759] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.094612] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b8c2ac-b789-435f-aec2-0fccd39f82db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.117391] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.118810] env[65107]: WARNING neutronclient.v2_0.client [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.119445] env[65107]: WARNING openstack [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.119801] env[65107]: WARNING openstack [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.127944] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4fc5f1f-baa1-4cea-936b-1ca1df6d55e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.149282] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 926.149282] env[65107]: value = "task-5103074" [ 926.149282] env[65107]: _type = "Task" [ 926.149282] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.158805] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103074, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.167861] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103073, 'name': Rename_Task, 'duration_secs': 0.187788} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.168144] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 926.168402] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab9a375c-8460-4f19-9bb3-3ce8bc974d11 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.175810] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 926.175810] env[65107]: value = "task-5103075" [ 926.175810] env[65107]: _type = "Task" [ 926.175810] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.184653] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.224473] env[65107]: DEBUG oslo_concurrency.lockutils [req-f62983be-7a13-418e-bf89-6904466e07ce req-5b83ea98-d9a3-41b0-8105-b4523090b54e service nova] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.231522] env[65107]: DEBUG nova.network.neutron [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Updated VIF entry in instance network info cache for port b58f1ae3-f99e-46a9-81be-e3d4925f3f85. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 926.231687] env[65107]: DEBUG nova.network.neutron [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Updating instance_info_cache with network_info: [{"id": "b58f1ae3-f99e-46a9-81be-e3d4925f3f85", "address": "fa:16:3e:b0:83:30", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb58f1ae3-f9", "ovs_interfaceid": "b58f1ae3-f99e-46a9-81be-e3d4925f3f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 926.472874] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b27c2b-0863-b31f-adcc-3c5e843f8fc9, 'name': SearchDatastore_Task, 'duration_secs': 0.010201} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.473186] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.473448] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.473805] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e54b8732-db1d-4858-8bee-a43fae9ddfbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.481969] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 926.481969] env[65107]: value = "task-5103076" [ 926.481969] env[65107]: _type = "Task" [ 926.481969] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.490717] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.561756] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "refresh_cache-70172f4f-2658-4bc1-857f-0dffde2d5ab0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.561756] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquired lock "refresh_cache-70172f4f-2658-4bc1-857f-0dffde2d5ab0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 926.561756] env[65107]: DEBUG nova.network.neutron [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 926.599503] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 52bd7ecb-4f53-443e-8916-3ff50dfab2b6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 926.599503] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 926.599753] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=100GB used_disk=18GB total_vcpus=48 used_vcpus=18 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '17', 'num_vm_resized': '1', 'num_task_None': '12', 'num_os_type_None': '17', 'num_proj_f379144b78764fe394039d87b043a946': '1', 'io_workload': '3', 'num_vm_active': '11', 'num_proj_e1fd7913632a4a5b9e076b5b156fa5fa': '1', 'num_proj_544c649453844164bac53c98b342543f': '2', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '1', 'num_proj_b981c5f6b3b64476a2406cd9549f7d6b': '1', 'num_proj_f883a2e25da34a548b020184acb9245e': '1', 'num_vm_rescued': '2', 'num_task_unrescuing': '1', 'num_proj_bfb7fca6907c4c32a1aa6b8023c39afb': '2', 'num_proj_f7db97cf40794c5196e31ab5fa8a0477': '1', 'num_proj_6f9188f3e35e4d06a977e2180918c616': '1', 'num_task_rebuild_spawning': '1', 'num_proj_198e62b5f04942f2b5a15eb2c7553094': '1', 'num_task_deleting': '1', 'num_proj_05918ba13682405ba5e5e7ebde13b42b': '1', 'num_proj_dced4f9aef2f49cf990203b693533aa1': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_152868ee0723449baac10f6a8ceeb738': '1', 'num_proj_c805fe68d0fd4147b9b4b7fdc9d6a6d0': '1', 'num_proj_75f92ee058014228898c5050f167f31c': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 926.668896] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.688365] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103075, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.735135] env[65107]: DEBUG oslo_concurrency.lockutils [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] Releasing lock "refresh_cache-fead9fe9-23f4-4d13-bf44-be0727057ddd" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.735549] env[65107]: DEBUG nova.compute.manager [req-25608ac4-29e2-4f25-95bc-064a6913fe52 req-036a47f9-a65c-4aff-891e-d2183ff3cfbb service nova] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Received event network-vif-deleted-3bd7ea5b-4f81-4cde-8a97-65795f27663e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 926.959963] env[65107]: DEBUG nova.compute.manager [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Received event network-vif-plugged-205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 926.961355] env[65107]: DEBUG oslo_concurrency.lockutils [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Acquiring lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.961471] env[65107]: DEBUG oslo_concurrency.lockutils [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.961953] env[65107]: DEBUG oslo_concurrency.lockutils [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.961953] env[65107]: DEBUG nova.compute.manager [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] No waiting events found dispatching network-vif-plugged-205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 926.961953] env[65107]: WARNING nova.compute.manager [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Received unexpected event network-vif-plugged-205cebba-e381-4ab7-a1f5-c04561c1b7a3 for instance with vm_state building and task_state spawning. [ 926.962981] env[65107]: DEBUG nova.compute.manager [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Received event network-changed-205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 926.962981] env[65107]: DEBUG nova.compute.manager [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Refreshing instance network info cache due to event network-changed-205cebba-e381-4ab7-a1f5-c04561c1b7a3. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 926.962981] env[65107]: DEBUG oslo_concurrency.lockutils [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Acquiring lock "refresh_cache-70172f4f-2658-4bc1-857f-0dffde2d5ab0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.997399] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103076, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.068894] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.070146] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.144552] env[65107]: DEBUG nova.network.neutron [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 927.166573] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da423ca-2f43-4efd-ac56-8b33f4c0d393 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.171991] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.173684] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.174102] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.188638] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d375046-b228-4c35-9de9-7a5d092b7503 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.195646] env[65107]: DEBUG oslo_vmware.api [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103075, 'name': PowerOnVM_Task, 'duration_secs': 0.563542} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.196439] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 927.196679] env[65107]: INFO nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Took 8.29 seconds to spawn the instance on the hypervisor. [ 927.197341] env[65107]: DEBUG nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 927.198339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75276a61-7a28-42c0-a378-18436799f49b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.233820] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132e0d95-e8ec-47af-9afc-20a4fd5f1b07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.247291] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1e6c76-fc64-4392-a0b8-dc5a2b48de49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.261931] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.283135] env[65107]: WARNING neutronclient.v2_0.client [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.283962] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.284532] env[65107]: WARNING openstack [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.404332] env[65107]: DEBUG nova.network.neutron [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Updating instance_info_cache with network_info: [{"id": "205cebba-e381-4ab7-a1f5-c04561c1b7a3", "address": "fa:16:3e:5c:a4:db", "network": {"id": "f4f68afc-c92e-445e-ac96-90e6d36de09e", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-792758191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f92ee058014228898c5050f167f31c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap205cebba-e3", "ovs_interfaceid": "205cebba-e381-4ab7-a1f5-c04561c1b7a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 927.494671] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635282} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.495050] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.495269] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.495539] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-123da64d-58be-445b-ab6f-637f1f6978f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.503725] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 927.503725] env[65107]: value = "task-5103077" [ 927.503725] env[65107]: _type = "Task" [ 927.503725] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.513353] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.664601] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103074, 'name': ReconfigVM_Task, 'duration_secs': 1.077193} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.664939] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be/4a42392a-a4b4-43b7-a410-0b69441435be.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.665610] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a1faa78-1971-449f-85a6-52b0434063b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.673983] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 927.673983] env[65107]: value = "task-5103078" [ 927.673983] env[65107]: _type = "Task" [ 927.673983] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.684562] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103078, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.753398] env[65107]: INFO nova.compute.manager [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Took 29.48 seconds to build instance. [ 927.765875] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 927.906869] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Releasing lock "refresh_cache-70172f4f-2658-4bc1-857f-0dffde2d5ab0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 927.907371] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Instance network_info: |[{"id": "205cebba-e381-4ab7-a1f5-c04561c1b7a3", "address": "fa:16:3e:5c:a4:db", "network": {"id": "f4f68afc-c92e-445e-ac96-90e6d36de09e", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-792758191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f92ee058014228898c5050f167f31c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap205cebba-e3", "ovs_interfaceid": "205cebba-e381-4ab7-a1f5-c04561c1b7a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 927.907742] env[65107]: DEBUG oslo_concurrency.lockutils [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Acquired lock "refresh_cache-70172f4f-2658-4bc1-857f-0dffde2d5ab0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.907945] env[65107]: DEBUG nova.network.neutron [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Refreshing network info cache for port 205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 927.909237] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:a4:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3952eb02-1162-48ed-8227-9c138960d583', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '205cebba-e381-4ab7-a1f5-c04561c1b7a3', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.917637] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Creating folder: Project (75f92ee058014228898c5050f167f31c). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 927.919363] env[65107]: WARNING neutronclient.v2_0.client [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.920374] env[65107]: WARNING openstack [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.920946] env[65107]: WARNING openstack [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.931757] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2616f665-3d7d-479a-9b4b-d769241cb4a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.946276] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Created folder: Project (75f92ee058014228898c5050f167f31c) in parent group-v992574. [ 927.946621] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Creating folder: Instances. Parent ref: group-v992790. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 927.946994] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf7a032b-7a56-40e8-a3d1-d86cd0b38b56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.958312] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Created folder: Instances in parent group-v992790. [ 927.958590] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 927.959602] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 927.959602] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c0bde6f-d288-43ea-9a19-1ac39658c65c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.980611] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.980611] env[65107]: value = "task-5103081" [ 927.980611] env[65107]: _type = "Task" [ 927.980611] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.989623] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103081, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.015638] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070358} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.015638] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.015923] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cd157c-b17f-43e1-a131-e697ece7dd91 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.039974] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.045674] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-249a0df8-e6bd-4eb5-bfaa-46984a1df3f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.072409] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 928.072409] env[65107]: value = "task-5103082" [ 928.072409] env[65107]: _type = "Task" [ 928.072409] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.086837] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.088528] env[65107]: WARNING openstack [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.089019] env[65107]: WARNING openstack [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.180710] env[65107]: WARNING neutronclient.v2_0.client [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.181894] env[65107]: WARNING openstack [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.181998] env[65107]: WARNING openstack [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.196315] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103078, 'name': Rename_Task, 'duration_secs': 0.462577} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.196627] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.196911] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36e2fdfa-580d-4c08-a0ed-29b8d57e78bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.207220] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 928.207220] env[65107]: value = "task-5103083" [ 928.207220] env[65107]: _type = "Task" [ 928.207220] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.216820] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.257114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-97b9ce9c-4102-4f70-bb1b-b7dada035584 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.260226] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.260501] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.270805] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 928.270805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.775s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.274891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.234s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.274891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.274891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.443s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.274891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.277830] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 20.570s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.302926] env[65107]: INFO nova.scheduler.client.report [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Deleted allocations for instance fb4229c6-b80d-49e9-a68b-baa12c0004ed [ 928.307023] env[65107]: INFO nova.scheduler.client.report [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Deleted allocations for instance 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e [ 928.312930] env[65107]: DEBUG nova.network.neutron [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Updated VIF entry in instance network info cache for port 205cebba-e381-4ab7-a1f5-c04561c1b7a3. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 928.314405] env[65107]: DEBUG nova.network.neutron [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Updating instance_info_cache with network_info: [{"id": "205cebba-e381-4ab7-a1f5-c04561c1b7a3", "address": "fa:16:3e:5c:a4:db", "network": {"id": "f4f68afc-c92e-445e-ac96-90e6d36de09e", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-792758191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f92ee058014228898c5050f167f31c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap205cebba-e3", "ovs_interfaceid": "205cebba-e381-4ab7-a1f5-c04561c1b7a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 928.492809] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103081, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.584123] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103082, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.719253] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103083, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.763639] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 928.822482] env[65107]: DEBUG oslo_concurrency.lockutils [req-31168026-789e-45ca-a2eb-b14b11734a69 req-a3746132-8ba3-4738-a832-1e21fd7864dc service nova] Releasing lock "refresh_cache-70172f4f-2658-4bc1-857f-0dffde2d5ab0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.824948] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b53a5ec1-c04a-4d93-aee0-ffd1fc16028d tempest-ServerGroupTestJSON-1803090293 tempest-ServerGroupTestJSON-1803090293-project-member] Lock "fb4229c6-b80d-49e9-a68b-baa12c0004ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.621s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.826028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-706917ab-2b8b-4485-bd6c-1591c5b4f70f tempest-ServersTestMultiNic-160748370 tempest-ServersTestMultiNic-160748370-project-member] Lock "64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.590s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.993617] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103081, 'name': CreateVM_Task, 'duration_secs': 0.544954} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.998023] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.998652] env[65107]: DEBUG nova.compute.manager [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-changed-e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 928.998850] env[65107]: DEBUG nova.compute.manager [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing instance network info cache due to event network-changed-e5280424-9c40-4f31-a39e-260899363c1d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 928.999053] env[65107]: DEBUG oslo_concurrency.lockutils [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.999206] env[65107]: DEBUG oslo_concurrency.lockutils [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 928.999366] env[65107]: DEBUG nova.network.neutron [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing network info cache for port e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 929.001368] env[65107]: WARNING neutronclient.v2_0.client [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.001368] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.001589] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.001740] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 929.002528] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01e535f4-5ec6-46a0-b225-600fea5228a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.010327] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 929.010327] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52220c60-429e-d2bb-6214-83aedda19072" [ 929.010327] env[65107]: _type = "Task" [ 929.010327] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.024230] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52220c60-429e-d2bb-6214-83aedda19072, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.085705] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103082, 'name': ReconfigVM_Task, 'duration_secs': 0.534425} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.086026] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Reconfigured VM instance instance-0000004a to attach disk [datastore1] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.086670] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a5a4e78-9420-4d6d-95a2-280ddfe92f50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.094649] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 929.094649] env[65107]: value = "task-5103084" [ 929.094649] env[65107]: _type = "Task" [ 929.094649] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.114785] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103084, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.218966] env[65107]: DEBUG oslo_vmware.api [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103083, 'name': PowerOnVM_Task, 'duration_secs': 0.944463} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.219036] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.219248] env[65107]: DEBUG nova.compute.manager [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 929.220086] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419b2ce9-53f1-4cb1-a9aa-37df0877956e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.252538] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64639613-c4e2-4c02-bc45-9b00b15cf6e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.264595] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a2986f-70e0-42ba-af83-2a066243f827 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.301523] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 929.302413] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b785924-8446-477d-a468-60f62673c30f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.314553] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad56c8d-2606-4f82-8a51-65e40e519ade {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.337452] env[65107]: DEBUG nova.compute.provider_tree [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.504740] env[65107]: WARNING neutronclient.v2_0.client [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.505508] env[65107]: WARNING openstack [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.505865] env[65107]: WARNING openstack [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.525553] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52220c60-429e-d2bb-6214-83aedda19072, 'name': SearchDatastore_Task, 'duration_secs': 0.01743} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.525902] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.526151] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.526398] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.526542] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.526725] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.527015] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54c79ecb-cc58-4b90-8af5-b1867e7b6d31 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.537826] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.537826] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.538530] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-813d3864-309a-4017-b4d2-f0b095bdce7c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.547799] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 929.547799] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e92633-5133-e2a5-f25e-5bea7e722fe1" [ 929.547799] env[65107]: _type = "Task" [ 929.547799] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.554217] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e92633-5133-e2a5-f25e-5bea7e722fe1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.605313] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103084, 'name': Rename_Task, 'duration_secs': 0.252319} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.605619] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.606098] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33f06aff-21c6-4a7e-ae0f-62646a63c48b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.613460] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 929.613460] env[65107]: value = "task-5103085" [ 929.613460] env[65107]: _type = "Task" [ 929.613460] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.624217] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103085, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.739164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 929.840418] env[65107]: DEBUG nova.scheduler.client.report [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.061257] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e92633-5133-e2a5-f25e-5bea7e722fe1, 'name': SearchDatastore_Task, 'duration_secs': 0.035576} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.062031] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae31a2c-e97e-4788-907f-f49f7cd83fb2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.067247] env[65107]: WARNING openstack [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.067645] env[65107]: WARNING openstack [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.080093] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 930.080093] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524ffdc7-e672-e099-29e6-db40ae1fac4e" [ 930.080093] env[65107]: _type = "Task" [ 930.080093] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.089902] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524ffdc7-e672-e099-29e6-db40ae1fac4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.124501] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103085, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.198083] env[65107]: WARNING neutronclient.v2_0.client [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.198780] env[65107]: WARNING openstack [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.199153] env[65107]: WARNING openstack [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.343338] env[65107]: DEBUG nova.network.neutron [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updated VIF entry in instance network info cache for port e5280424-9c40-4f31-a39e-260899363c1d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 930.343338] env[65107]: DEBUG nova.network.neutron [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.592979] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524ffdc7-e672-e099-29e6-db40ae1fac4e, 'name': SearchDatastore_Task, 'duration_secs': 0.011637} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.593260] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 930.593773] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 70172f4f-2658-4bc1-857f-0dffde2d5ab0/70172f4f-2658-4bc1-857f-0dffde2d5ab0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.593942] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-768410d8-4aa5-42ea-831c-20ad2aba5d85 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.602656] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 930.602656] env[65107]: value = "task-5103086" [ 930.602656] env[65107]: _type = "Task" [ 930.602656] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.612930] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103086, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.625659] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103085, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.846916] env[65107]: DEBUG oslo_concurrency.lockutils [req-dc5545a1-55e7-40ea-99de-c657e88d4f0f req-7b1b730c-08eb-4e75-a098-393ffd07d55a service nova] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 930.857161] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.579s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.857161] env[65107]: DEBUG nova.compute.manager [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=65107) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5433}} [ 930.862951] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.418s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.862951] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.863424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.403s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.864928] env[65107]: INFO nova.compute.claims [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.889601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "4a42392a-a4b4-43b7-a410-0b69441435be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.889657] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "4a42392a-a4b4-43b7-a410-0b69441435be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.889877] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "4a42392a-a4b4-43b7-a410-0b69441435be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.890076] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "4a42392a-a4b4-43b7-a410-0b69441435be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.890261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "4a42392a-a4b4-43b7-a410-0b69441435be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.893555] env[65107]: INFO nova.compute.manager [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Terminating instance [ 930.901961] env[65107]: INFO nova.scheduler.client.report [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Deleted allocations for instance fdb5603d-24b0-4fdd-b8cc-25c0ebee5311 [ 931.114129] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103086, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.137532] env[65107]: DEBUG oslo_vmware.api [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103085, 'name': PowerOnVM_Task, 'duration_secs': 1.03534} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.137957] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.138205] env[65107]: INFO nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Took 8.31 seconds to spawn the instance on the hypervisor. [ 931.138389] env[65107]: DEBUG nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 931.139285] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d20ee4-5320-4237-9c9f-eac5414b7c1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.399331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "refresh_cache-4a42392a-a4b4-43b7-a410-0b69441435be" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.399514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquired lock "refresh_cache-4a42392a-a4b4-43b7-a410-0b69441435be" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.399690] env[65107]: DEBUG nova.network.neutron [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 931.410485] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc1183fc-f653-4f4b-920a-7c7858bfb0b0 tempest-MigrationsAdminTest-673773847 tempest-MigrationsAdminTest-673773847-project-member] Lock "fdb5603d-24b0-4fdd-b8cc-25c0ebee5311" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.152s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.427184] env[65107]: INFO nova.scheduler.client.report [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted allocation for migration cb52d55b-9ce2-4ccf-996d-1ceb56e3a62b [ 931.615720] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103086, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641143} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.616051] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 70172f4f-2658-4bc1-857f-0dffde2d5ab0/70172f4f-2658-4bc1-857f-0dffde2d5ab0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.616343] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.616639] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e7c94f1c-ad95-44b8-bf36-5cdbbb5d9676 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.624900] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 931.624900] env[65107]: value = "task-5103087" [ 931.624900] env[65107]: _type = "Task" [ 931.624900] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.636796] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103087, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.660951] env[65107]: INFO nova.compute.manager [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Took 30.64 seconds to build instance. [ 931.904328] env[65107]: WARNING neutronclient.v2_0.client [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 931.904775] env[65107]: WARNING openstack [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.905167] env[65107]: WARNING openstack [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 931.936335] env[65107]: DEBUG nova.network.neutron [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 931.938906] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bc104a49-8efb-431c-92de-0372ff743f90 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 27.603s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.044992] env[65107]: DEBUG nova.network.neutron [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 932.143042] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103087, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086723} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.143369] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.144666] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77064c0c-4eb6-49b0-94b3-b28939e91ef4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.172466] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 70172f4f-2658-4bc1-857f-0dffde2d5ab0/70172f4f-2658-4bc1-857f-0dffde2d5ab0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.175977] env[65107]: DEBUG oslo_concurrency.lockutils [None req-53dd4a18-8746-4b87-badc-a512350cdcb1 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.166s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.176322] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57cb6557-b56b-426a-920b-8d4981bfd913 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.199533] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 932.199533] env[65107]: value = "task-5103088" [ 932.199533] env[65107]: _type = "Task" [ 932.199533] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.209417] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103088, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.396579] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666949df-4ca2-4589-8c3f-3d0d95876a4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.406020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38dcf58-6248-43d5-95c2-465bb4e43849 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.438869] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d1654c-4188-4ec0-8e9f-3d62f4571cfd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.448520] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae4dbd5-e51f-4bee-9cee-c1091ffc92ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.464127] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.548514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Releasing lock "refresh_cache-4a42392a-a4b4-43b7-a410-0b69441435be" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.549114] env[65107]: DEBUG nova.compute.manager [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 932.549416] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.550802] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c329e179-5ffa-4d98-9805-481159081ae2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.561071] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.561550] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa4aad5e-69f0-43bd-87cd-cd0f93aff4ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.569272] env[65107]: DEBUG oslo_vmware.api [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 932.569272] env[65107]: value = "task-5103089" [ 932.569272] env[65107]: _type = "Task" [ 932.569272] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.580103] env[65107]: DEBUG oslo_vmware.api [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.714248] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103088, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.990252] env[65107]: ERROR nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [req-2b9cd622-c019-4802-a679-2f86ac6a6647] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2b9cd622-c019-4802-a679-2f86ac6a6647"}]} [ 933.009854] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 933.029693] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 933.029919] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.042809] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 933.073322] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 933.088088] env[65107]: DEBUG oslo_vmware.api [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103089, 'name': PowerOffVM_Task, 'duration_secs': 0.18401} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.088900] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.088900] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.089050] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4bc0360-866c-48bb-8983-ffb687227f84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.106648] env[65107]: DEBUG nova.objects.instance [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'flavor' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.120252] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.120252] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.120252] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Deleting the datastore file [datastore1] 4a42392a-a4b4-43b7-a410-0b69441435be {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.120505] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-016938fd-44fe-4132-9263-9074d3697d87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.135187] env[65107]: DEBUG oslo_vmware.api [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for the task: (returnval){ [ 933.135187] env[65107]: value = "task-5103091" [ 933.135187] env[65107]: _type = "Task" [ 933.135187] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.148594] env[65107]: DEBUG oslo_vmware.api [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.214454] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103088, 'name': ReconfigVM_Task, 'duration_secs': 0.528736} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.214625] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 70172f4f-2658-4bc1-857f-0dffde2d5ab0/70172f4f-2658-4bc1-857f-0dffde2d5ab0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.215284] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0588d83b-04b4-4ada-a1f4-7562a2d536fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.223743] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 933.223743] env[65107]: value = "task-5103092" [ 933.223743] env[65107]: _type = "Task" [ 933.223743] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.234538] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103092, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.534869] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80c3a85-571b-4963-ba7d-16de47a759c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.544902] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5112b9-61b9-4f81-a529-c3ccd21b1709 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.578395] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e5bf1e-c16e-4e87-b836-635cb1a2e394 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.586735] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ebb4e5-8dbd-43b4-8419-594fd0085933 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.601292] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.612832] env[65107]: DEBUG oslo_concurrency.lockutils [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.613042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 933.613250] env[65107]: DEBUG nova.network.neutron [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 933.613443] env[65107]: DEBUG nova.objects.instance [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'info_cache' on Instance uuid 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.646591] env[65107]: DEBUG oslo_vmware.api [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Task: {'id': task-5103091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160929} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.646867] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.647154] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.647245] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.647404] env[65107]: INFO nova.compute.manager [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Took 1.10 seconds to destroy the instance on the hypervisor. [ 933.647707] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 933.647907] env[65107]: DEBUG nova.compute.manager [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 933.648032] env[65107]: DEBUG nova.network.neutron [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 933.648385] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.648932] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.649209] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.674180] env[65107]: DEBUG nova.network.neutron [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 933.674180] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.736037] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103092, 'name': Rename_Task, 'duration_secs': 0.162876} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.736037] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.736803] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75b636b5-2a23-49a5-90c3-d7d68cb365be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.744112] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 933.744112] env[65107]: value = "task-5103093" [ 933.744112] env[65107]: _type = "Task" [ 933.744112] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.754330] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.849670] env[65107]: INFO nova.compute.manager [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Rebuilding instance [ 933.892872] env[65107]: DEBUG nova.compute.manager [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 933.893739] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c139723d-59fc-4f52-9250-06dd3aa38f1d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.118188] env[65107]: DEBUG nova.objects.base [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Object Instance<71e6e280-50c3-4655-a6a3-40ea4c650d96> lazy-loaded attributes: flavor,info_cache {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 934.123059] env[65107]: ERROR nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [req-4b10ffe0-dfb8-4399-b07e-249473c19cf5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4b10ffe0-dfb8-4399-b07e-249473c19cf5"}]} [ 934.142906] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 934.159414] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 934.159730] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.173658] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 934.175896] env[65107]: DEBUG nova.network.neutron [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 934.194953] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 934.260326] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103093, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.583926] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "c5fb4b1c-5540-4645-b996-f397729beb23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.584227] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "c5fb4b1c-5540-4645-b996-f397729beb23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.621534] env[65107]: WARNING neutronclient.v2_0.client [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.622601] env[65107]: WARNING openstack [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.622896] env[65107]: WARNING openstack [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.633090] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d27d1c-2521-4159-a9b3-5fe201bd05bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.642170] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c94a52-f9cd-4b1f-b18b-71c50c547dbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.678118] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6a119b-87f7-4879-9e69-baa1e023caef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.681147] env[65107]: INFO nova.compute.manager [-] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Took 1.03 seconds to deallocate network for instance. [ 934.695320] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531fd7ca-e8de-4737-9bb2-25144120dca3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.712427] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.758202] env[65107]: DEBUG oslo_vmware.api [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103093, 'name': PowerOnVM_Task, 'duration_secs': 0.51762} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.758401] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.758601] env[65107]: INFO nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Took 9.20 seconds to spawn the instance on the hypervisor. [ 934.758777] env[65107]: DEBUG nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 934.759588] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58fe22a-2e6d-4b0e-a036-d9828ccbd849 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.784388] env[65107]: WARNING openstack [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.784708] env[65107]: WARNING openstack [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.865275] env[65107]: WARNING neutronclient.v2_0.client [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.866040] env[65107]: WARNING openstack [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.866386] env[65107]: WARNING openstack [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.910847] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.911638] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d4c088e-efd4-42cc-8d82-0210edf9e8a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.919971] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 934.919971] env[65107]: value = "task-5103094" [ 934.919971] env[65107]: _type = "Task" [ 934.919971] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.934677] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.964804] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "58345821-536a-46ad-af55-92de7552e924" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.966751] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "58345821-536a-46ad-af55-92de7552e924" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.989997] env[65107]: DEBUG nova.network.neutron [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [{"id": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "address": "fa:16:3e:99:b8:ff", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d38d576-18", "ovs_interfaceid": "3d38d576-1871-49a5-bc10-aef1a37abfa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 935.089999] env[65107]: DEBUG nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 935.195590] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.236892] env[65107]: ERROR nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [req-3a7d5d3a-8c06-4f15-8f58-db64d875f8a4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3a7d5d3a-8c06-4f15-8f58-db64d875f8a4"}]} [ 935.255038] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 935.269382] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 935.269772] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.279844] env[65107]: INFO nova.compute.manager [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Took 33.81 seconds to build instance. [ 935.289203] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 935.311949] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 935.433497] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103094, 'name': PowerOffVM_Task, 'duration_secs': 0.218616} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.433786] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.434018] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.435252] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73c8511-1ed3-44b8-9971-31bab97e90ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.444795] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.445135] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4406f4d9-73f0-41b7-a9aa-9d41cb06de8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.469183] env[65107]: DEBUG nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 935.495236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-71e6e280-50c3-4655-a6a3-40ea4c650d96" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 935.520317] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.520539] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.520714] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore1] fead9fe9-23f4-4d13-bf44-be0727057ddd {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.521014] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edbc156e-eedd-451e-98bf-7393feebba8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.536661] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 935.536661] env[65107]: value = "task-5103096" [ 935.536661] env[65107]: _type = "Task" [ 935.536661] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.550539] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.612591] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.731845] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8e838b-5b03-4ec1-839b-0d9c0d7818f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.740354] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e8bf92-12ff-4589-a9f8-ee92a63b72fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.774185] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d89f437-389e-460e-a46f-33788091bc06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.781785] env[65107]: DEBUG oslo_concurrency.lockutils [None req-43248bfd-13ac-4bef-ad2e-c9b6c43d5936 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.321s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 935.784509] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dc33fa-f789-4467-82fb-1169d0d5bedc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.802134] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.993724] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.058687] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138624} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.058906] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.059245] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 936.059338] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 936.179591] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.179875] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.180122] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.180308] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.180480] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.184692] env[65107]: INFO nova.compute.manager [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Terminating instance [ 936.343613] env[65107]: DEBUG nova.scheduler.client.report [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 936.343932] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 111 to 112 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 936.344140] env[65107]: DEBUG nova.compute.provider_tree [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 936.500382] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.500677] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67bc4435-2b53-4d01-a32b-621439859134 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.511416] env[65107]: DEBUG oslo_vmware.api [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 936.511416] env[65107]: value = "task-5103097" [ 936.511416] env[65107]: _type = "Task" [ 936.511416] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.521920] env[65107]: DEBUG oslo_vmware.api [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.693320] env[65107]: DEBUG nova.compute.manager [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 936.693577] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 936.694599] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b74890-76f9-4e13-84e7-3f06f7237840 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.706097] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.706607] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91a1605d-553f-4e79-a072-90c41bb5f8e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.716560] env[65107]: DEBUG oslo_vmware.api [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 936.716560] env[65107]: value = "task-5103098" [ 936.716560] env[65107]: _type = "Task" [ 936.716560] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.728753] env[65107]: DEBUG oslo_vmware.api [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.850093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.986s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.851303] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 936.854299] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.529s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.856177] env[65107]: INFO nova.compute.claims [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.023202] env[65107]: DEBUG oslo_vmware.api [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103097, 'name': PowerOnVM_Task, 'duration_secs': 0.451578} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.024515] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.024515] env[65107]: DEBUG nova.compute.manager [None req-debcd220-2fc8-458e-b508-8fbb30aa7e08 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 937.024843] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1aca2c-3e59-4cda-8ce1-49b0250b34fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.106231] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 937.106733] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 937.110020] env[65107]: DEBUG nova.virt.hardware [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 937.110020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f52245-384f-4b67-832b-a792419b5f28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.120849] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05c469e-ea82-41e5-8fad-4ce50c701295 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.139500] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:83:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b58f1ae3-f99e-46a9-81be-e3d4925f3f85', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.147280] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 937.147745] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.148098] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8db42f5-45b2-4538-9ec7-5cd66cfff64d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.169086] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.169086] env[65107]: value = "task-5103099" [ 937.169086] env[65107]: _type = "Task" [ 937.169086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.181923] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103099, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.230481] env[65107]: DEBUG oslo_vmware.api [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103098, 'name': PowerOffVM_Task, 'duration_secs': 0.229798} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.230481] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.230481] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.230733] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5881e06f-e151-4a5e-b7e4-8dc6a264bf2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.312192] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 937.312459] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 937.312646] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Deleting the datastore file [datastore2] 70172f4f-2658-4bc1-857f-0dffde2d5ab0 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.313099] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29edded5-29de-4517-980e-d70f2364df18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.322264] env[65107]: DEBUG oslo_vmware.api [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for the task: (returnval){ [ 937.322264] env[65107]: value = "task-5103101" [ 937.322264] env[65107]: _type = "Task" [ 937.322264] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.331609] env[65107]: DEBUG oslo_vmware.api [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.364011] env[65107]: DEBUG nova.compute.utils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 937.368662] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 937.368709] env[65107]: DEBUG nova.network.neutron [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 937.369111] env[65107]: WARNING neutronclient.v2_0.client [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.369627] env[65107]: WARNING neutronclient.v2_0.client [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.370996] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.370996] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.459134] env[65107]: DEBUG nova.policy [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0480b6c310e5483c9b5eb3b9b6e972bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4daf9bbfa3e9477d83d9d8c3d1fb5a02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 937.688453] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103099, 'name': CreateVM_Task, 'duration_secs': 0.416934} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.688453] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 937.688453] env[65107]: WARNING neutronclient.v2_0.client [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 937.688453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.688453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.688699] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 937.691882] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9fc4a50-65bc-468f-8e3f-f130ce8c3474 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.699756] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 937.699756] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5279f333-7fd2-92b5-a276-eea4c3259f7b" [ 937.699756] env[65107]: _type = "Task" [ 937.699756] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.711551] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5279f333-7fd2-92b5-a276-eea4c3259f7b, 'name': SearchDatastore_Task, 'duration_secs': 0.010177} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.711862] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 937.712107] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.712348] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.712486] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.712683] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.712939] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8671d8cc-cb23-49f2-a655-d59d85698b25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.725287] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.725287] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.726264] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e493692-c3e4-4c37-8f1f-50f8bf18743c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.733768] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 937.733768] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528f3a3f-31a3-ad0d-926c-fc5e5245a568" [ 937.733768] env[65107]: _type = "Task" [ 937.733768] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.744285] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528f3a3f-31a3-ad0d-926c-fc5e5245a568, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.833882] env[65107]: DEBUG oslo_vmware.api [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Task: {'id': task-5103101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18507} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.834958] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.834958] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.834958] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.834958] env[65107]: INFO nova.compute.manager [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 937.834958] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 937.835280] env[65107]: DEBUG nova.compute.manager [-] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 937.835316] env[65107]: DEBUG nova.network.neutron [-] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 937.835670] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.839222] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.839222] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.882168] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 937.889092] env[65107]: DEBUG nova.network.neutron [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Successfully created port: dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 937.921239] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.923338] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.923471] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.923669] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.923848] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.924015] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.926684] env[65107]: INFO nova.compute.manager [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Terminating instance [ 938.187633] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.187957] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.188273] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.188273] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.188424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.190577] env[65107]: INFO nova.compute.manager [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Terminating instance [ 938.247689] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528f3a3f-31a3-ad0d-926c-fc5e5245a568, 'name': SearchDatastore_Task, 'duration_secs': 0.011233} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.248889] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24f3fb7b-c00e-4de1-a784-fc2661b381ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.258501] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 938.258501] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52536026-7247-9ac2-88a5-79476499c58b" [ 938.258501] env[65107]: _type = "Task" [ 938.258501] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.271828] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52536026-7247-9ac2-88a5-79476499c58b, 'name': SearchDatastore_Task, 'duration_secs': 0.010431} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.272169] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 938.272465] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.272733] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32b14e2a-9bc1-498b-aee0-d125b52ad531 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.283648] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 938.283648] env[65107]: value = "task-5103102" [ 938.283648] env[65107]: _type = "Task" [ 938.283648] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.292603] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103102, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.370766] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23354d9-6799-454b-a2a7-5e6eb9951c99 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.379375] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc9a6ea-e0cd-420d-8e86-609d340ed242 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.418985] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccb376b-f245-4042-a2e3-65ad69358e87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.430103] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff330a8-24f6-479d-887e-76c687b77b32 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.439940] env[65107]: DEBUG nova.compute.manager [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 938.439940] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.441713] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399400ed-dbf6-4542-bd07-afcc0653eab9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.459470] env[65107]: DEBUG nova.compute.provider_tree [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 938.463187] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.463657] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a354c9f-a6e1-4883-9266-40d6477b9253 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.472702] env[65107]: DEBUG oslo_vmware.api [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 938.472702] env[65107]: value = "task-5103103" [ 938.472702] env[65107]: _type = "Task" [ 938.472702] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.482788] env[65107]: DEBUG oslo_vmware.api [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.695933] env[65107]: DEBUG nova.compute.manager [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 938.697071] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.697613] env[65107]: DEBUG nova.network.neutron [-] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 938.701473] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41a6224-b310-4970-a647-f3f6c9b53a6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.713206] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.714040] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af8ec30c-25f6-473b-b160-17d40b242415 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.727068] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 938.727068] env[65107]: value = "task-5103104" [ 938.727068] env[65107]: _type = "Task" [ 938.727068] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.739971] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5103104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.802503] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103102, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.842349] env[65107]: DEBUG nova.compute.manager [req-cb109d9d-2f84-4e54-8ebc-ce89cc52e7de req-71a74405-5308-44f3-b33e-72ec63f1e9c7 service nova] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Received event network-vif-deleted-205cebba-e381-4ab7-a1f5-c04561c1b7a3 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 938.920413] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 938.957828] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 938.958138] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 938.958301] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 938.958477] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 938.958626] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 938.958862] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 938.958987] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.959183] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 938.959348] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 938.959545] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 938.959709] env[65107]: DEBUG nova.virt.hardware [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 938.960627] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ec4bff-7675-47b4-9d63-7cf887335fe5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.982546] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ce8c06-0318-45d0-9926-d879d253f729 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.005015] env[65107]: DEBUG oslo_vmware.api [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103103, 'name': PowerOffVM_Task, 'duration_secs': 0.244139} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.005384] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.005588] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.006037] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8433be9-a3a4-48b0-ae0e-0d4d4a0a49ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.013517] env[65107]: ERROR nova.scheduler.client.report [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [req-69d4f64e-a9cd-43e3-8c24-fa646d0b6d34] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-69d4f64e-a9cd-43e3-8c24-fa646d0b6d34"}]} [ 939.035797] env[65107]: DEBUG nova.scheduler.client.report [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 939.057713] env[65107]: DEBUG nova.scheduler.client.report [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 939.057842] env[65107]: DEBUG nova.compute.provider_tree [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.074062] env[65107]: DEBUG nova.scheduler.client.report [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 939.084213] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.084598] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.085376] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleting the datastore file [datastore1] 71e6e280-50c3-4655-a6a3-40ea4c650d96 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.085376] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-985ec0cd-4e1e-4c04-a7e5-6a5ca8097443 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.094030] env[65107]: DEBUG oslo_vmware.api [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 939.094030] env[65107]: value = "task-5103106" [ 939.094030] env[65107]: _type = "Task" [ 939.094030] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.100395] env[65107]: DEBUG nova.scheduler.client.report [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 939.106994] env[65107]: DEBUG oslo_vmware.api [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.208090] env[65107]: INFO nova.compute.manager [-] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Took 1.37 seconds to deallocate network for instance. [ 939.241796] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5103104, 'name': PowerOffVM_Task, 'duration_secs': 0.308309} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.242298] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.242595] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.243031] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0442b5b-bf23-41ae-a5ed-421d68d11fbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.294947] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103102, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537798} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.295396] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.295514] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.295781] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e2a85d3-3ac6-4b22-921e-b757b9640fbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.306652] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 939.306652] env[65107]: value = "task-5103108" [ 939.306652] env[65107]: _type = "Task" [ 939.306652] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.318413] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103108, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.335620] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.336109] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.336247] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Deleting the datastore file [datastore1] cddb14db-34c3-4783-b21e-e0b4623a3a9e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.336467] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e45ff4ac-a318-4710-9d5b-6f91a5a80554 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.348530] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for the task: (returnval){ [ 939.348530] env[65107]: value = "task-5103109" [ 939.348530] env[65107]: _type = "Task" [ 939.348530] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.358322] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5103109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.435893] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "f4b1352b-1c55-4987-a298-69431c4e565d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.436285] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "f4b1352b-1c55-4987-a298-69431c4e565d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.490644] env[65107]: DEBUG nova.compute.manager [req-dbf43996-13ea-4144-8b9b-118c1ea8fcc9 req-8aeba7fc-8c3c-4f7a-9bbe-68227d6dc5cb service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Received event network-vif-plugged-dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 939.490867] env[65107]: DEBUG oslo_concurrency.lockutils [req-dbf43996-13ea-4144-8b9b-118c1ea8fcc9 req-8aeba7fc-8c3c-4f7a-9bbe-68227d6dc5cb service nova] Acquiring lock "888d60d4-bc9c-48af-97f9-434883884e8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.491260] env[65107]: DEBUG oslo_concurrency.lockutils [req-dbf43996-13ea-4144-8b9b-118c1ea8fcc9 req-8aeba7fc-8c3c-4f7a-9bbe-68227d6dc5cb service nova] Lock "888d60d4-bc9c-48af-97f9-434883884e8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.491503] env[65107]: DEBUG oslo_concurrency.lockutils [req-dbf43996-13ea-4144-8b9b-118c1ea8fcc9 req-8aeba7fc-8c3c-4f7a-9bbe-68227d6dc5cb service nova] Lock "888d60d4-bc9c-48af-97f9-434883884e8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.491789] env[65107]: DEBUG nova.compute.manager [req-dbf43996-13ea-4144-8b9b-118c1ea8fcc9 req-8aeba7fc-8c3c-4f7a-9bbe-68227d6dc5cb service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] No waiting events found dispatching network-vif-plugged-dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 939.491930] env[65107]: WARNING nova.compute.manager [req-dbf43996-13ea-4144-8b9b-118c1ea8fcc9 req-8aeba7fc-8c3c-4f7a-9bbe-68227d6dc5cb service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Received unexpected event network-vif-plugged-dacbb269-1799-4b10-8007-7778dddae890 for instance with vm_state building and task_state spawning. [ 939.595427] env[65107]: DEBUG nova.network.neutron [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Successfully updated port: dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 939.617953] env[65107]: DEBUG oslo_vmware.api [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281699} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.618968] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.619174] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.619353] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.619717] env[65107]: INFO nova.compute.manager [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 1.18 seconds to destroy the instance on the hypervisor. [ 939.620916] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 939.623804] env[65107]: DEBUG nova.compute.manager [-] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 939.623983] env[65107]: DEBUG nova.network.neutron [-] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 939.624287] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.624939] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.625241] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.664023] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc91c04-5260-4972-a2d3-e842256a3bb6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.668084] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.676972] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a7c59-51b2-4dbe-88f5-9acaba309bd8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.737186] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.739179] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ab11a5-046a-4d20-b77d-0a75ebd7db19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.751672] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb891963-33c4-4676-afde-a5924a462749 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.776363] env[65107]: DEBUG nova.compute.provider_tree [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.817997] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103108, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.224122} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.818748] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.819449] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30690a42-bf9f-4b4b-87eb-5ab595f5dbce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.845179] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.845475] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4103de33-e207-4207-9909-9c0e3235fc62 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.871388] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5103109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.873227] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 939.873227] env[65107]: value = "task-5103110" [ 939.873227] env[65107]: _type = "Task" [ 939.873227] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.882978] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103110, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.942380] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 940.106718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "refresh_cache-888d60d4-bc9c-48af-97f9-434883884e8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.106718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired lock "refresh_cache-888d60d4-bc9c-48af-97f9-434883884e8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.106718] env[65107]: DEBUG nova.network.neutron [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 940.327900] env[65107]: DEBUG nova.scheduler.client.report [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 113 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 940.328197] env[65107]: DEBUG nova.compute.provider_tree [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 113 to 114 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 940.328372] env[65107]: DEBUG nova.compute.provider_tree [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.375326] env[65107]: DEBUG oslo_vmware.api [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Task: {'id': task-5103109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.561906} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.382427] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.382985] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.384030] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.384030] env[65107]: INFO nova.compute.manager [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Took 1.69 seconds to destroy the instance on the hypervisor. [ 940.384562] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 940.388024] env[65107]: DEBUG nova.compute.manager [-] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 940.388024] env[65107]: DEBUG nova.network.neutron [-] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 940.388024] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.388024] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.388024] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.407924] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.467974] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.479717] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.610450] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.610860] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.656540] env[65107]: DEBUG nova.network.neutron [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 940.675166] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.675592] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.719161] env[65107]: DEBUG nova.network.neutron [-] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 940.760018] env[65107]: WARNING neutronclient.v2_0.client [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.760018] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.760018] env[65107]: WARNING openstack [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.833897] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.980s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.834571] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 940.837713] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.474s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.837980] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.840820] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.368s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.841097] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.843031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.373s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.843105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.844983] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.367s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.845269] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.847155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.811s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.848933] env[65107]: INFO nova.compute.claims [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.859634] env[65107]: DEBUG nova.network.neutron [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Updating instance_info_cache with network_info: [{"id": "dacbb269-1799-4b10-8007-7778dddae890", "address": "fa:16:3e:60:c5:21", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdacbb269-17", "ovs_interfaceid": "dacbb269-1799-4b10-8007-7778dddae890", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 940.869856] env[65107]: INFO nova.scheduler.client.report [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Deleted allocations for instance 1a813da6-f777-4888-b2ae-fbad58e01f61 [ 940.876335] env[65107]: INFO nova.scheduler.client.report [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Deleted allocations for instance 88cac920-3c1e-41dc-9e52-f51f48a80f56 [ 940.893018] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103110, 'name': ReconfigVM_Task, 'duration_secs': 0.913279} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.895353] env[65107]: INFO nova.scheduler.client.report [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleted allocations for instance 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c [ 940.896478] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Reconfigured VM instance instance-0000004a to attach disk [datastore2] fead9fe9-23f4-4d13-bf44-be0727057ddd/fead9fe9-23f4-4d13-bf44-be0727057ddd.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.902650] env[65107]: INFO nova.scheduler.client.report [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Deleted allocations for instance 9fa949ea-358c-46e7-a0f2-4c3275493b64 [ 940.903586] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c176a539-b1da-444c-80f1-2a88136bad5c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.917015] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 940.917015] env[65107]: value = "task-5103111" [ 940.917015] env[65107]: _type = "Task" [ 940.917015] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.933449] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103111, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.033673] env[65107]: DEBUG nova.compute.manager [req-c1112838-f065-4d82-851f-1876949542b7 req-3bfe3914-199f-447d-a52d-d0433a2b4512 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Received event network-vif-deleted-b3279b33-9092-457f-88cc-4c2cc95e6cdc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 941.033890] env[65107]: INFO nova.compute.manager [req-c1112838-f065-4d82-851f-1876949542b7 req-3bfe3914-199f-447d-a52d-d0433a2b4512 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Neutron deleted interface b3279b33-9092-457f-88cc-4c2cc95e6cdc; detaching it from the instance and deleting it from the info cache [ 941.034022] env[65107]: DEBUG nova.network.neutron [req-c1112838-f065-4d82-851f-1876949542b7 req-3bfe3914-199f-447d-a52d-d0433a2b4512 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 941.222862] env[65107]: INFO nova.compute.manager [-] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Took 1.60 seconds to deallocate network for instance. [ 941.354060] env[65107]: DEBUG nova.compute.utils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 941.359196] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 941.359196] env[65107]: DEBUG nova.network.neutron [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 941.359196] env[65107]: WARNING neutronclient.v2_0.client [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.359196] env[65107]: WARNING neutronclient.v2_0.client [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.359403] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.359838] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.374169] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Releasing lock "refresh_cache-888d60d4-bc9c-48af-97f9-434883884e8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.374169] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Instance network_info: |[{"id": "dacbb269-1799-4b10-8007-7778dddae890", "address": "fa:16:3e:60:c5:21", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdacbb269-17", "ovs_interfaceid": "dacbb269-1799-4b10-8007-7778dddae890", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 941.374169] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:c5:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dacbb269-1799-4b10-8007-7778dddae890', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.380945] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Creating folder: Project (4daf9bbfa3e9477d83d9d8c3d1fb5a02). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 941.384985] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cf4167f-df31-4d19-971b-5ecd92d70134 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.388235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be0cc03b-44bb-4f24-a164-d0b893755b1b tempest-ServerExternalEventsTest-2032441286 tempest-ServerExternalEventsTest-2032441286-project-member] Lock "1a813da6-f777-4888-b2ae-fbad58e01f61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.341s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.393222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1289560-a386-47fe-a2da-9b7eee48b661 tempest-ServerMetadataTestJSON-1796680905 tempest-ServerMetadataTestJSON-1796680905-project-member] Lock "88cac920-3c1e-41dc-9e52-f51f48a80f56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.972s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.403188] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Created folder: Project (4daf9bbfa3e9477d83d9d8c3d1fb5a02) in parent group-v992574. [ 941.403488] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Creating folder: Instances. Parent ref: group-v992794. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 941.403707] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bd79860-9ad2-4ead-b951-36e462ce6fd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.415095] env[65107]: DEBUG nova.policy [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ccab6477dd44878f78a9d6428d3cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4778fe9152224fd29d1f6220a19b5a36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 941.423611] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6c6372d8-7a98-4913-a1d8-ae3acd661af2 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "21c471cd-2435-4e4e-82cc-8cdf03f1fa6c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.260s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.424519] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Created folder: Instances in parent group-v992794. [ 941.424767] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 941.426340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-733e38f7-d410-4a1a-ad73-82fe4f0c5427 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "9fa949ea-358c-46e7-a0f2-4c3275493b64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.538s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.429316] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.429949] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6100cac5-fb45-416d-ab93-e5fd493c81ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.454517] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103111, 'name': Rename_Task, 'duration_secs': 0.159172} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.457331] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.457331] env[65107]: DEBUG nova.network.neutron [-] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 941.458407] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.458407] env[65107]: value = "task-5103114" [ 941.458407] env[65107]: _type = "Task" [ 941.458407] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.458909] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b09bde35-0f0f-44f8-8a98-b2b2a3ee24af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.472648] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103114, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.475437] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 941.475437] env[65107]: value = "task-5103115" [ 941.475437] env[65107]: _type = "Task" [ 941.475437] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.486058] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.537753] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7069e371-3ec8-48e7-87d4-d2c01b67613e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.550381] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe32a5b-3ee4-4b07-a714-6cc87d0f30f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.605445] env[65107]: DEBUG nova.compute.manager [req-c1112838-f065-4d82-851f-1876949542b7 req-3bfe3914-199f-447d-a52d-d0433a2b4512 service nova] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Detach interface failed, port_id=b3279b33-9092-457f-88cc-4c2cc95e6cdc, reason: Instance cddb14db-34c3-4783-b21e-e0b4623a3a9e could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 941.730914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.812073] env[65107]: DEBUG nova.network.neutron [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Successfully created port: 1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 941.858680] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 941.959316] env[65107]: INFO nova.compute.manager [-] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Took 1.57 seconds to deallocate network for instance. [ 941.976755] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103114, 'name': CreateVM_Task, 'duration_secs': 0.413115} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.983737] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.985163] env[65107]: WARNING neutronclient.v2_0.client [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.985617] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.985922] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.986126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 941.986876] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dd56363-9aae-4153-9674-48eb9f0e14b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.993497] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103115, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.998045] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 941.998045] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527fbdb0-1e7d-8176-4dc4-74cfe54aabc2" [ 941.998045] env[65107]: _type = "Task" [ 941.998045] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.015115] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527fbdb0-1e7d-8176-4dc4-74cfe54aabc2, 'name': SearchDatastore_Task, 'duration_secs': 0.012004} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.018377] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.018720] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.018973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.019133] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.019315] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.021015] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33d9527b-26ce-4cdb-a04d-4e636e8b04b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.030720] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.030924] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.036031] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e788274a-3928-4f0d-baca-28ca5218d232 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.042031] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 942.042031] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5218dbd9-0869-42f2-1848-010933813211" [ 942.042031] env[65107]: _type = "Task" [ 942.042031] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.050860] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5218dbd9-0869-42f2-1848-010933813211, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.308872] env[65107]: DEBUG nova.compute.manager [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Received event network-changed-dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 942.308950] env[65107]: DEBUG nova.compute.manager [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Refreshing instance network info cache due to event network-changed-dacbb269-1799-4b10-8007-7778dddae890. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 942.309262] env[65107]: DEBUG oslo_concurrency.lockutils [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Acquiring lock "refresh_cache-888d60d4-bc9c-48af-97f9-434883884e8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.309405] env[65107]: DEBUG oslo_concurrency.lockutils [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Acquired lock "refresh_cache-888d60d4-bc9c-48af-97f9-434883884e8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.309564] env[65107]: DEBUG nova.network.neutron [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Refreshing network info cache for port dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 942.348437] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6aa269f-d971-4c26-a2e7-74170e125e2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.356445] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca607ed-007e-4bcd-aeee-ffbd1d0d1fe2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.394757] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ae6ef6-e2fd-462a-b12b-3f4a5e4da9ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.402884] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d978ab-f9e8-449a-80d8-b12eaf1c1c45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.418578] env[65107]: DEBUG nova.compute.provider_tree [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.471301] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.487639] env[65107]: DEBUG oslo_vmware.api [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103115, 'name': PowerOnVM_Task, 'duration_secs': 0.554317} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.487945] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.488070] env[65107]: DEBUG nova.compute.manager [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 942.488827] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccaa174-f812-4a52-a970-4a3918993422 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.551712] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5218dbd9-0869-42f2-1848-010933813211, 'name': SearchDatastore_Task, 'duration_secs': 0.014964} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.552591] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-785cda5a-040b-47a3-9cc8-84c7bbd232af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.559839] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 942.559839] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5248dffb-81f6-8a8a-9973-85be7861646f" [ 942.559839] env[65107]: _type = "Task" [ 942.559839] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.571914] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5248dffb-81f6-8a8a-9973-85be7861646f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.812705] env[65107]: WARNING neutronclient.v2_0.client [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.813723] env[65107]: WARNING openstack [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.815971] env[65107]: WARNING openstack [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.896858] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 942.923499] env[65107]: DEBUG nova.scheduler.client.report [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 942.938401] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 942.938644] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 942.938818] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 942.939645] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 942.939645] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 942.939645] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 942.939645] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.939645] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 942.940076] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 942.940076] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 942.940162] env[65107]: DEBUG nova.virt.hardware [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 942.941016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d00b601-00d1-48bd-831b-a8ac5793f719 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.951324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df4edbd-93af-4c94-9a95-72f6687cd7e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.010306] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.071312] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5248dffb-81f6-8a8a-9973-85be7861646f, 'name': SearchDatastore_Task, 'duration_secs': 0.044944} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.071651] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.071956] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 888d60d4-bc9c-48af-97f9-434883884e8a/888d60d4-bc9c-48af-97f9-434883884e8a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.072267] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c0a76f4-8c24-4d57-9379-542da03173f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.080234] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 943.080234] env[65107]: value = "task-5103116" [ 943.080234] env[65107]: _type = "Task" [ 943.080234] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.090630] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.433413] env[65107]: WARNING openstack [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.433903] env[65107]: WARNING openstack [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.447269] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.447269] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 943.450080] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.883s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.452240] env[65107]: INFO nova.compute.claims [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.503503] env[65107]: DEBUG nova.network.neutron [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Successfully updated port: 1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 943.595265] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103116, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.646802] env[65107]: WARNING neutronclient.v2_0.client [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.647610] env[65107]: WARNING openstack [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.648011] env[65107]: WARNING openstack [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.809024] env[65107]: DEBUG nova.network.neutron [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Updated VIF entry in instance network info cache for port dacbb269-1799-4b10-8007-7778dddae890. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 943.809024] env[65107]: DEBUG nova.network.neutron [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Updating instance_info_cache with network_info: [{"id": "dacbb269-1799-4b10-8007-7778dddae890", "address": "fa:16:3e:60:c5:21", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdacbb269-17", "ovs_interfaceid": "dacbb269-1799-4b10-8007-7778dddae890", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 943.959488] env[65107]: DEBUG nova.compute.utils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 943.961244] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 943.961464] env[65107]: DEBUG nova.network.neutron [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 943.961903] env[65107]: WARNING neutronclient.v2_0.client [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.962334] env[65107]: WARNING neutronclient.v2_0.client [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.962983] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.963409] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.006243] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.006460] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.006645] env[65107]: DEBUG nova.network.neutron [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 944.011424] env[65107]: DEBUG nova.policy [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '002065730b3949da9b9f44ba5c40c378', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa38929b22f24f4fa948c5dad97d6e51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 944.017834] env[65107]: DEBUG nova.compute.manager [req-5643f0b4-57a1-4975-8328-06c2da03f0e4 req-bab382ad-4378-4963-b3c6-5003ef8605fc service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Received event network-vif-plugged-1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 944.018223] env[65107]: DEBUG oslo_concurrency.lockutils [req-5643f0b4-57a1-4975-8328-06c2da03f0e4 req-bab382ad-4378-4963-b3c6-5003ef8605fc service nova] Acquiring lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.018352] env[65107]: DEBUG oslo_concurrency.lockutils [req-5643f0b4-57a1-4975-8328-06c2da03f0e4 req-bab382ad-4378-4963-b3c6-5003ef8605fc service nova] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.018541] env[65107]: DEBUG oslo_concurrency.lockutils [req-5643f0b4-57a1-4975-8328-06c2da03f0e4 req-bab382ad-4378-4963-b3c6-5003ef8605fc service nova] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.018810] env[65107]: DEBUG nova.compute.manager [req-5643f0b4-57a1-4975-8328-06c2da03f0e4 req-bab382ad-4378-4963-b3c6-5003ef8605fc service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] No waiting events found dispatching network-vif-plugged-1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 944.019217] env[65107]: WARNING nova.compute.manager [req-5643f0b4-57a1-4975-8328-06c2da03f0e4 req-bab382ad-4378-4963-b3c6-5003ef8605fc service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Received unexpected event network-vif-plugged-1c2c4d44-fde7-4882-9b77-f18455fd3786 for instance with vm_state building and task_state spawning. [ 944.095830] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103116, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.842278} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.095830] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 888d60d4-bc9c-48af-97f9-434883884e8a/888d60d4-bc9c-48af-97f9-434883884e8a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.095830] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.096090] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09a39af3-c90c-4662-8276-df9234e87035 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.104173] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 944.104173] env[65107]: value = "task-5103117" [ 944.104173] env[65107]: _type = "Task" [ 944.104173] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.116049] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103117, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.311881] env[65107]: DEBUG oslo_concurrency.lockutils [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] Releasing lock "refresh_cache-888d60d4-bc9c-48af-97f9-434883884e8a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.313779] env[65107]: DEBUG nova.compute.manager [req-7380b6b9-9017-4ba5-a626-6aa02a8d663f req-34e14552-ed68-4a8f-8913-66a36e759194 service nova] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Received event network-vif-deleted-3d38d576-1871-49a5-bc10-aef1a37abfa0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 944.387712] env[65107]: DEBUG nova.network.neutron [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Successfully created port: c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 944.471631] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 944.481718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "40474981-eeef-492e-8a8b-aaea7f554c02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.481977] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.482231] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "40474981-eeef-492e-8a8b-aaea7f554c02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.482422] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.482590] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.484872] env[65107]: INFO nova.compute.manager [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Terminating instance [ 944.515979] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.516398] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.598853] env[65107]: DEBUG nova.network.neutron [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 944.617457] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103117, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075389} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.617678] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.618954] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3872e639-2a8f-4306-b4d3-77577291c81e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.644375] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 888d60d4-bc9c-48af-97f9-434883884e8a/888d60d4-bc9c-48af-97f9-434883884e8a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.645937] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.646391] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.655409] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-570f0bf1-17d9-4272-9738-5295b7b1c615 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.679083] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 944.679083] env[65107]: value = "task-5103118" [ 944.679083] env[65107]: _type = "Task" [ 944.679083] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.691767] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103118, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.762977] env[65107]: WARNING neutronclient.v2_0.client [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.763681] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.764144] env[65107]: WARNING openstack [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.858786] env[65107]: DEBUG nova.network.neutron [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Updating instance_info_cache with network_info: [{"id": "1c2c4d44-fde7-4882-9b77-f18455fd3786", "address": "fa:16:3e:80:1b:df", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c2c4d44-fd", "ovs_interfaceid": "1c2c4d44-fde7-4882-9b77-f18455fd3786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 944.965482] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1e6b41-1e84-4109-b8de-056cb2c94176 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.977026] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5d57be-a160-47bf-a4f6-beed0bc9da29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.990977] env[65107]: DEBUG nova.compute.manager [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 944.991474] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.019522] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8064023-5939-48ba-a61b-69e091ab08cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.025055] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29ff38d-3b14-4913-98b5-60da2fb720c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.036182] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4514caf7-6a38-4259-8921-35155ddb12f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.041338] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.042116] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f46fc0fc-b7c6-40bb-b0aa-9f8b3dd9ef51 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.056691] env[65107]: DEBUG nova.compute.provider_tree [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.060694] env[65107]: DEBUG oslo_vmware.api [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 945.060694] env[65107]: value = "task-5103119" [ 945.060694] env[65107]: _type = "Task" [ 945.060694] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.073140] env[65107]: DEBUG oslo_vmware.api [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103119, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.191529] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103118, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.367283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.367683] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Instance network_info: |[{"id": "1c2c4d44-fde7-4882-9b77-f18455fd3786", "address": "fa:16:3e:80:1b:df", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c2c4d44-fd", "ovs_interfaceid": "1c2c4d44-fde7-4882-9b77-f18455fd3786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 945.368181] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:1b:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c2c4d44-fde7-4882-9b77-f18455fd3786', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.376538] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 945.376952] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "fead9fe9-23f4-4d13-bf44-be0727057ddd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.377207] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.377408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "fead9fe9-23f4-4d13-bf44-be0727057ddd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.377592] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.377759] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.379491] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.379776] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e84276b-6a27-4976-b658-4f444b1823db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.395720] env[65107]: INFO nova.compute.manager [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Terminating instance [ 945.403827] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.403827] env[65107]: value = "task-5103120" [ 945.403827] env[65107]: _type = "Task" [ 945.403827] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.417555] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103120, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.488459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "13c96e2f-46de-46e0-a505-60dfec9e95ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.488735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "13c96e2f-46de-46e0-a505-60dfec9e95ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.490840] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 945.514743] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 945.515068] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 945.515245] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 945.515453] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 945.515618] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 945.515798] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 945.515984] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.516160] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 945.516388] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 945.516539] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 945.516653] env[65107]: DEBUG nova.virt.hardware [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 945.517595] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4bf5d7-be9e-4106-a7c9-31f34e162a7a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.532640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3455a8c-3f9d-45df-aae2-ba9f4075505f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.562915] env[65107]: DEBUG nova.scheduler.client.report [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 945.575952] env[65107]: DEBUG oslo_vmware.api [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103119, 'name': PowerOffVM_Task, 'duration_secs': 0.236729} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.576312] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.576403] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.576652] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92e3174f-6ffd-4c0e-a33a-43451814be0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.647445] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.647668] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.647847] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleting the datastore file [datastore1] 40474981-eeef-492e-8a8b-aaea7f554c02 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.648147] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f65c439-a616-497f-beba-2bd72d8a5dc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.656923] env[65107]: DEBUG oslo_vmware.api [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for the task: (returnval){ [ 945.656923] env[65107]: value = "task-5103122" [ 945.656923] env[65107]: _type = "Task" [ 945.656923] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.664771] env[65107]: DEBUG oslo_vmware.api [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.691941] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103118, 'name': ReconfigVM_Task, 'duration_secs': 0.619634} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.692438] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 888d60d4-bc9c-48af-97f9-434883884e8a/888d60d4-bc9c-48af-97f9-434883884e8a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.693519] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20d2ade6-1503-465b-a140-d674cb22843b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.701483] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 945.701483] env[65107]: value = "task-5103123" [ 945.701483] env[65107]: _type = "Task" [ 945.701483] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.713037] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103123, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.900692] env[65107]: DEBUG nova.compute.manager [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 945.901552] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.904027] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2786ce7f-ea4a-4966-a0c3-b0203ee87f24 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.916385] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103120, 'name': CreateVM_Task, 'duration_secs': 0.499663} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.919728] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.919728] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.919728] env[65107]: WARNING neutronclient.v2_0.client [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.920169] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.921389] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.921389] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 945.921389] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85e3fbdd-5a9c-4805-a74e-e99e7a9662c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.922915] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42e5d339-1ca3-4c68-a495-b47276bec570 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.928404] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 945.928404] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52242591-2cbc-b822-1a2a-7a49ea76a7d4" [ 945.928404] env[65107]: _type = "Task" [ 945.928404] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.933522] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 945.933522] env[65107]: value = "task-5103124" [ 945.933522] env[65107]: _type = "Task" [ 945.933522] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.940283] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52242591-2cbc-b822-1a2a-7a49ea76a7d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.947874] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.994646] env[65107]: DEBUG nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 946.018553] env[65107]: DEBUG nova.network.neutron [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Successfully updated port: c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 946.071691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.072594] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 946.075199] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.605s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.076797] env[65107]: DEBUG nova.objects.instance [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lazy-loading 'resources' on Instance uuid 3a975793-f480-4f54-85c9-2c2aea11c18a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.167224] env[65107]: DEBUG oslo_vmware.api [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Task: {'id': task-5103122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.295347} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.167644] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.167856] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.168058] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.168235] env[65107]: INFO nova.compute.manager [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Took 1.18 seconds to destroy the instance on the hypervisor. [ 946.168478] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 946.168726] env[65107]: DEBUG nova.compute.manager [-] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 946.168854] env[65107]: DEBUG nova.network.neutron [-] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 946.172284] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.172284] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.172284] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.213355] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103123, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.233952] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.291266] env[65107]: DEBUG nova.compute.manager [req-20201b2d-7c5c-4779-bd7a-173768e9b93b req-8a1cfcc0-3af7-4a40-960c-3e09daa58a4b service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Received event network-vif-plugged-c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 946.291472] env[65107]: DEBUG oslo_concurrency.lockutils [req-20201b2d-7c5c-4779-bd7a-173768e9b93b req-8a1cfcc0-3af7-4a40-960c-3e09daa58a4b service nova] Acquiring lock "d5766a03-054f-40ea-a57e-e640664ca683-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.291677] env[65107]: DEBUG oslo_concurrency.lockutils [req-20201b2d-7c5c-4779-bd7a-173768e9b93b req-8a1cfcc0-3af7-4a40-960c-3e09daa58a4b service nova] Lock "d5766a03-054f-40ea-a57e-e640664ca683-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.291838] env[65107]: DEBUG oslo_concurrency.lockutils [req-20201b2d-7c5c-4779-bd7a-173768e9b93b req-8a1cfcc0-3af7-4a40-960c-3e09daa58a4b service nova] Lock "d5766a03-054f-40ea-a57e-e640664ca683-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.292085] env[65107]: DEBUG nova.compute.manager [req-20201b2d-7c5c-4779-bd7a-173768e9b93b req-8a1cfcc0-3af7-4a40-960c-3e09daa58a4b service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] No waiting events found dispatching network-vif-plugged-c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 946.292681] env[65107]: WARNING nova.compute.manager [req-20201b2d-7c5c-4779-bd7a-173768e9b93b req-8a1cfcc0-3af7-4a40-960c-3e09daa58a4b service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Received unexpected event network-vif-plugged-c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 for instance with vm_state building and task_state spawning. [ 946.451265] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52242591-2cbc-b822-1a2a-7a49ea76a7d4, 'name': SearchDatastore_Task, 'duration_secs': 0.020603} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.453412] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.453664] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.454224] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.454224] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.454408] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.454585] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.454780] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b7b7b39-f0af-4d14-839c-75c7b5b23d5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.467583] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.467776] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.468649] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dab08856-a066-4090-91b9-a830e67163e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.476456] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 946.476456] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528bab03-67b3-df20-418a-a727ff8a7152" [ 946.476456] env[65107]: _type = "Task" [ 946.476456] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.488927] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528bab03-67b3-df20-418a-a727ff8a7152, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.491835] env[65107]: DEBUG nova.compute.manager [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Received event network-changed-1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 946.492093] env[65107]: DEBUG nova.compute.manager [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Refreshing instance network info cache due to event network-changed-1c2c4d44-fde7-4882-9b77-f18455fd3786. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 946.492385] env[65107]: DEBUG oslo_concurrency.lockutils [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Acquiring lock "refresh_cache-23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.492781] env[65107]: DEBUG oslo_concurrency.lockutils [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Acquired lock "refresh_cache-23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.493031] env[65107]: DEBUG nova.network.neutron [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Refreshing network info cache for port 1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 946.508025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.508150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.522437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "refresh_cache-d5766a03-054f-40ea-a57e-e640664ca683" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.522437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquired lock "refresh_cache-d5766a03-054f-40ea-a57e-e640664ca683" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.522437] env[65107]: DEBUG nova.network.neutron [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 946.527073] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.579049] env[65107]: DEBUG nova.compute.utils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 946.586039] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 946.586305] env[65107]: DEBUG nova.network.neutron [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 946.586774] env[65107]: WARNING neutronclient.v2_0.client [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.587212] env[65107]: WARNING neutronclient.v2_0.client [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.587885] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.588265] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.636988] env[65107]: DEBUG nova.policy [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32155078a41249a297470fd4083860df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f9188f3e35e4d06a977e2180918c616', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 946.715346] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103123, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.949974] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.965364] env[65107]: DEBUG nova.network.neutron [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Successfully created port: 853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 946.967744] env[65107]: DEBUG nova.network.neutron [-] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 946.987194] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528bab03-67b3-df20-418a-a727ff8a7152, 'name': SearchDatastore_Task, 'duration_secs': 0.011462} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.988206] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb06b302-fa6b-4f44-969f-df43027b0075 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.000240] env[65107]: WARNING neutronclient.v2_0.client [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.000240] env[65107]: WARNING openstack [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.000240] env[65107]: WARNING openstack [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.009615] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 947.009615] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f29ed6-1495-16b3-e6ab-eb1b4a8e8852" [ 947.009615] env[65107]: _type = "Task" [ 947.009615] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.011200] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 947.014252] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fbb1fa-1bc2-4980-8841-42caf4115377 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.030027] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.030027] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.036679] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f29ed6-1495-16b3-e6ab-eb1b4a8e8852, 'name': SearchDatastore_Task, 'duration_secs': 0.010413} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.040545] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.040651] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151/23c3b100-30ac-44f9-8e2d-d8e3f2f4e151.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.040986] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30c98559-7f10-4de1-9795-8a3d9a2682e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.044589] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75cc874-cf01-4150-a19f-2229677a23e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.914750] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 947.914750] env[65107]: INFO nova.compute.manager [-] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Took 1.75 seconds to deallocate network for instance. [ 947.921072] env[65107]: DEBUG nova.network.neutron [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 947.935967] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a59db9-5cc4-468c-8f6d-6e1a851acf0e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.938890] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 947.938890] env[65107]: value = "task-5103125" [ 947.938890] env[65107]: _type = "Task" [ 947.938890] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.956076] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103124, 'name': PowerOffVM_Task, 'duration_secs': 1.172015} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.956193] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103123, 'name': Rename_Task, 'duration_secs': 1.211029} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.958028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.958968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c93010f-aaef-444a-b907-a103ffa2d3f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.963275] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 947.963455] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 947.964440] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.969019] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-931ec8f9-c1a5-47e7-b37c-ded1ce3a042a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.971113] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49f31827-1d0d-4e2e-bcc7-6d325fe5c9f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.971215] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103125, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49152} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.971978] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151/23c3b100-30ac-44f9-8e2d-d8e3f2f4e151.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.972153] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.973048] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b12fb3a-08fe-443d-acaa-142a1824f27a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.984431] env[65107]: DEBUG nova.compute.provider_tree [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.988020] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 947.988020] env[65107]: value = "task-5103127" [ 947.988020] env[65107]: _type = "Task" [ 947.988020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.993313] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 947.993313] env[65107]: value = "task-5103128" [ 947.993313] env[65107]: _type = "Task" [ 947.993313] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.997071] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103127, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.010204] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103128, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.052651] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.052968] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.053197] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore2] fead9fe9-23f4-4d13-bf44-be0727057ddd {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.053512] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d46b620-32b9-4b84-8052-fe9be4deaa94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.062235] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 948.062235] env[65107]: value = "task-5103129" [ 948.062235] env[65107]: _type = "Task" [ 948.062235] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.074195] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.307633] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.307883] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.362322] env[65107]: WARNING openstack [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.365798] env[65107]: WARNING openstack [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.442852] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3f34d6a2-0370-4239-9556-b4333e454d58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.444050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3f34d6a2-0370-4239-9556-b4333e454d58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.444050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3f34d6a2-0370-4239-9556-b4333e454d58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.445946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3f34d6a2-0370-4239-9556-b4333e454d58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.445946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3f34d6a2-0370-4239-9556-b4333e454d58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.447597] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.448222] env[65107]: INFO nova.compute.manager [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Terminating instance [ 948.456587] env[65107]: WARNING neutronclient.v2_0.client [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.457462] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.457852] env[65107]: WARNING openstack [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.489303] env[65107]: DEBUG nova.scheduler.client.report [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.498177] env[65107]: WARNING neutronclient.v2_0.client [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.500248] env[65107]: WARNING openstack [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.500248] env[65107]: WARNING openstack [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.520527] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103127, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.520527] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103128, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104651} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.520718] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.521692] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c77909c-7a18-4f95-99a9-97b022169b82 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.547332] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151/23c3b100-30ac-44f9-8e2d-d8e3f2f4e151.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.547332] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce9b69c5-8307-47e8-842a-d198140a4899 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.576113] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 948.576113] env[65107]: value = "task-5103130" [ 948.576113] env[65107]: _type = "Task" [ 948.576113] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.577681] env[65107]: DEBUG oslo_vmware.api [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210098} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.580962] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.581266] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.581366] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.581548] env[65107]: INFO nova.compute.manager [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Took 2.68 seconds to destroy the instance on the hypervisor. [ 948.582195] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 948.582195] env[65107]: DEBUG nova.compute.manager [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 948.582195] env[65107]: DEBUG nova.network.neutron [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 948.582428] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.582975] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.583252] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.598842] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103130, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.666528] env[65107]: DEBUG nova.network.neutron [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Updating instance_info_cache with network_info: [{"id": "c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6", "address": "fa:16:3e:1f:30:d0", "network": {"id": "5ee29ab4-c310-42f8-bcee-360908e767b0", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1296127457-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "aa38929b22f24f4fa948c5dad97d6e51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e399f8-2c", "ovs_interfaceid": "c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 948.749823] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.789771] env[65107]: DEBUG nova.network.neutron [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Updated VIF entry in instance network info cache for port 1c2c4d44-fde7-4882-9b77-f18455fd3786. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 948.790432] env[65107]: DEBUG nova.network.neutron [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Updating instance_info_cache with network_info: [{"id": "1c2c4d44-fde7-4882-9b77-f18455fd3786", "address": "fa:16:3e:80:1b:df", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c2c4d44-fd", "ovs_interfaceid": "1c2c4d44-fde7-4882-9b77-f18455fd3786", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 948.937246] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 948.953263] env[65107]: DEBUG nova.compute.manager [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 948.953511] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.954456] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15830bd4-e878-4ee8-82ab-5075a1ebb48e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.969230] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.973337] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c77f18a-f10c-48bc-b222-2f7738fb1bab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.983425] env[65107]: DEBUG oslo_vmware.api [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 948.983425] env[65107]: value = "task-5103131" [ 948.983425] env[65107]: _type = "Task" [ 948.983425] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.990528] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8080fdfc891c4c0724ea7e77bd63ceeb',container_format='bare',created_at=2025-12-12T13:26:18Z,direct_url=,disk_format='vmdk',id=57f71a90-320c-4d88-8671-5ad259d2a173,min_disk=1,min_ram=0,name='tempest-test-snap-1481872627',owner='6f9188f3e35e4d06a977e2180918c616',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-12-12T13:26:34Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 948.990870] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 948.991594] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 948.993630] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 948.993630] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 948.993630] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 948.993630] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.993630] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 948.993891] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 948.994076] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 948.994217] env[65107]: DEBUG nova.virt.hardware [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 948.995886] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdae0f3b-6cdb-43ee-af87-c57fa95df17f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.010641] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.935s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.017350] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.716s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.019044] env[65107]: INFO nova.compute.claims [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.023877] env[65107]: DEBUG oslo_vmware.api [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.031975] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51988c19-9551-4cc3-bb79-99284e544c23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.038787] env[65107]: DEBUG oslo_vmware.api [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103127, 'name': PowerOnVM_Task, 'duration_secs': 0.550039} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.038935] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.039127] env[65107]: INFO nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Took 10.12 seconds to spawn the instance on the hypervisor. [ 949.039399] env[65107]: DEBUG nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 949.040694] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2b7cd8-5250-4d89-91e4-d0e517a00a18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.045137] env[65107]: INFO nova.scheduler.client.report [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Deleted allocations for instance 3a975793-f480-4f54-85c9-2c2aea11c18a [ 949.088349] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103130, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.170566] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Releasing lock "refresh_cache-d5766a03-054f-40ea-a57e-e640664ca683" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.174023] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Instance network_info: |[{"id": "c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6", "address": "fa:16:3e:1f:30:d0", "network": {"id": "5ee29ab4-c310-42f8-bcee-360908e767b0", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1296127457-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "aa38929b22f24f4fa948c5dad97d6e51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e399f8-2c", "ovs_interfaceid": "c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 949.174023] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:30:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.181025] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Creating folder: Project (aa38929b22f24f4fa948c5dad97d6e51). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.181355] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a7faf51-1e01-48f3-9f2b-09f282dee17c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.184057] env[65107]: DEBUG nova.network.neutron [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Successfully updated port: 853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 949.198322] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Created folder: Project (aa38929b22f24f4fa948c5dad97d6e51) in parent group-v992574. [ 949.198695] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Creating folder: Instances. Parent ref: group-v992798. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.200533] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d37e3fc-305c-4d2a-ab55-5716b40ce4fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.212768] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Created folder: Instances in parent group-v992798. [ 949.213084] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 949.213663] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.213904] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8787e6a5-7916-4993-add2-31767bf29ad0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.234627] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.234627] env[65107]: value = "task-5103134" [ 949.234627] env[65107]: _type = "Task" [ 949.234627] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.245090] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103134, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.293520] env[65107]: DEBUG oslo_concurrency.lockutils [req-49e9d6b8-f1d3-4f53-92e1-f3c7f19b25eb req-ae705536-c680-4b14-b3d2-3e5034b257d6 service nova] Releasing lock "refresh_cache-23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.477993] env[65107]: DEBUG nova.compute.manager [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Received event network-changed-c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 949.478263] env[65107]: DEBUG nova.compute.manager [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Refreshing instance network info cache due to event network-changed-c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 949.478464] env[65107]: DEBUG oslo_concurrency.lockutils [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Acquiring lock "refresh_cache-d5766a03-054f-40ea-a57e-e640664ca683" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.478630] env[65107]: DEBUG oslo_concurrency.lockutils [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Acquired lock "refresh_cache-d5766a03-054f-40ea-a57e-e640664ca683" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.478777] env[65107]: DEBUG nova.network.neutron [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Refreshing network info cache for port c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 949.494410] env[65107]: DEBUG oslo_vmware.api [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103131, 'name': PowerOffVM_Task, 'duration_secs': 0.265502} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.494688] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.494970] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.495644] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bf2507e-80ef-432b-820d-601701e1cea1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.528567] env[65107]: DEBUG nova.compute.manager [req-4e2ff743-c415-4183-b8a8-63b7ae9b4994 req-d6f97c2d-2589-4575-bec5-001b1e76824d service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Received event network-vif-plugged-853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 949.529070] env[65107]: DEBUG oslo_concurrency.lockutils [req-4e2ff743-c415-4183-b8a8-63b7ae9b4994 req-d6f97c2d-2589-4575-bec5-001b1e76824d service nova] Acquiring lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.529262] env[65107]: DEBUG oslo_concurrency.lockutils [req-4e2ff743-c415-4183-b8a8-63b7ae9b4994 req-d6f97c2d-2589-4575-bec5-001b1e76824d service nova] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.529453] env[65107]: DEBUG oslo_concurrency.lockutils [req-4e2ff743-c415-4183-b8a8-63b7ae9b4994 req-d6f97c2d-2589-4575-bec5-001b1e76824d service nova] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.529623] env[65107]: DEBUG nova.compute.manager [req-4e2ff743-c415-4183-b8a8-63b7ae9b4994 req-d6f97c2d-2589-4575-bec5-001b1e76824d service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] No waiting events found dispatching network-vif-plugged-853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 949.529859] env[65107]: WARNING nova.compute.manager [req-4e2ff743-c415-4183-b8a8-63b7ae9b4994 req-d6f97c2d-2589-4575-bec5-001b1e76824d service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Received unexpected event network-vif-plugged-853b4d7e-c8ee-4570-a59c-310c21f06168 for instance with vm_state building and task_state spawning. [ 949.572353] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b98d868-dd77-48cd-bd45-6b6b1c16bd36 tempest-InstanceActionsV221TestJSON-2125704182 tempest-InstanceActionsV221TestJSON-2125704182-project-member] Lock "3a975793-f480-4f54-85c9-2c2aea11c18a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.561s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.582490] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.582707] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.583178] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleting the datastore file [datastore2] 3f34d6a2-0370-4239-9556-b4333e454d58 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.585568] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f04620af-2db9-4cd0-963e-0dd71ce13544 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.588182] env[65107]: INFO nova.compute.manager [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Took 40.15 seconds to build instance. [ 949.602200] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103130, 'name': ReconfigVM_Task, 'duration_secs': 0.913772} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.604037] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151/23c3b100-30ac-44f9-8e2d-d8e3f2f4e151.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.604856] env[65107]: DEBUG oslo_vmware.api [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 949.604856] env[65107]: value = "task-5103136" [ 949.604856] env[65107]: _type = "Task" [ 949.604856] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.605923] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49984ca8-ab14-4b55-acd5-3a91ac18d047 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.613676] env[65107]: DEBUG nova.network.neutron [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.622842] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 949.622842] env[65107]: value = "task-5103137" [ 949.622842] env[65107]: _type = "Task" [ 949.622842] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.623605] env[65107]: DEBUG oslo_vmware.api [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103136, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.636675] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103137, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.687223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "refresh_cache-52bd7ecb-4f53-443e-8916-3ff50dfab2b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.687599] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "refresh_cache-52bd7ecb-4f53-443e-8916-3ff50dfab2b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.687963] env[65107]: DEBUG nova.network.neutron [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 949.747817] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103134, 'name': CreateVM_Task, 'duration_secs': 0.400523} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.748035] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 949.748655] env[65107]: WARNING neutronclient.v2_0.client [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.749033] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.749180] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.749522] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 949.749907] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5b81f4c-1864-469d-a586-0fc1ce9bd4a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.755177] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 949.755177] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52940256-47b1-c891-1a26-911e1c6fb92e" [ 949.755177] env[65107]: _type = "Task" [ 949.755177] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.763963] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52940256-47b1-c891-1a26-911e1c6fb92e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.982054] env[65107]: WARNING neutronclient.v2_0.client [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.983349] env[65107]: WARNING openstack [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.983349] env[65107]: WARNING openstack [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.096834] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe8fbf0c-253f-4536-bb0e-6302d78a65f1 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "888d60d4-bc9c-48af-97f9-434883884e8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.674s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.116380] env[65107]: INFO nova.compute.manager [-] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Took 1.53 seconds to deallocate network for instance. [ 950.128631] env[65107]: DEBUG oslo_vmware.api [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103136, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193121} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.135909] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.136463] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.136463] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.137433] env[65107]: INFO nova.compute.manager [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Took 1.18 seconds to destroy the instance on the hypervisor. [ 950.137433] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 950.137433] env[65107]: DEBUG nova.compute.manager [-] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 950.137433] env[65107]: DEBUG nova.network.neutron [-] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 950.137628] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.138653] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.138653] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.155113] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103137, 'name': Rename_Task, 'duration_secs': 0.160236} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.155113] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.155113] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-936de803-9981-4544-a54e-305363bb17e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.161440] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 950.161440] env[65107]: value = "task-5103138" [ 950.161440] env[65107]: _type = "Task" [ 950.161440] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.171039] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.191848] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.192590] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.245235] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.268244] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52940256-47b1-c891-1a26-911e1c6fb92e, 'name': SearchDatastore_Task, 'duration_secs': 0.010057} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.271945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.271945] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.271945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.271945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.272148] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.273251] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37cbd1de-f01a-48fc-ac9e-ee59b2aae995 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.288672] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.288874] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.289658] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5e24475-06ab-4c54-9ae9-5efb86b7d1bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.296440] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 950.296440] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52570a77-22fc-5af2-b826-885c77c092dc" [ 950.296440] env[65107]: _type = "Task" [ 950.296440] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.309257] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52570a77-22fc-5af2-b826-885c77c092dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.373635] env[65107]: DEBUG nova.network.neutron [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 950.391731] env[65107]: WARNING openstack [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.392189] env[65107]: WARNING openstack [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.506815] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796b2ca7-1449-4ad6-b323-15b1ef837d52 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.516377] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1224abe-0ae5-408c-943b-6dc19f011f67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.556279] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13701554-04d3-4245-8817-d2a429e20529 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.565166] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafa7970-5ec6-4c66-9e42-3aaa6e6e48ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.580452] env[65107]: DEBUG nova.compute.provider_tree [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.635667] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.674834] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103138, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.810984] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52570a77-22fc-5af2-b826-885c77c092dc, 'name': SearchDatastore_Task, 'duration_secs': 0.012801} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.812114] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de32dcbf-3475-4248-b161-5a1c17b99303 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.819093] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 950.819093] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52579961-536f-e634-8271-766a1a1df578" [ 950.819093] env[65107]: _type = "Task" [ 950.819093] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.830774] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52579961-536f-e634-8271-766a1a1df578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.914931] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.915396] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.020438] env[65107]: WARNING neutronclient.v2_0.client [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.022314] env[65107]: WARNING openstack [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.022314] env[65107]: WARNING openstack [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.087144] env[65107]: DEBUG nova.scheduler.client.report [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 951.174694] env[65107]: DEBUG oslo_vmware.api [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103138, 'name': PowerOnVM_Task, 'duration_secs': 0.990311} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.175015] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.175232] env[65107]: INFO nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Took 8.28 seconds to spawn the instance on the hypervisor. [ 951.175928] env[65107]: DEBUG nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 951.176267] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec3e23a-22d5-48ee-802a-e6e31d84ac10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.187632] env[65107]: WARNING neutronclient.v2_0.client [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.187959] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.188450] env[65107]: WARNING openstack [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.247011] env[65107]: DEBUG nova.network.neutron [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Updated VIF entry in instance network info cache for port c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 951.248137] env[65107]: DEBUG nova.network.neutron [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Updating instance_info_cache with network_info: [{"id": "c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6", "address": "fa:16:3e:1f:30:d0", "network": {"id": "5ee29ab4-c310-42f8-bcee-360908e767b0", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1296127457-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "aa38929b22f24f4fa948c5dad97d6e51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e399f8-2c", "ovs_interfaceid": "c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.334282] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52579961-536f-e634-8271-766a1a1df578, 'name': SearchDatastore_Task, 'duration_secs': 0.023984} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.335295] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.335722] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] d5766a03-054f-40ea-a57e-e640664ca683/d5766a03-054f-40ea-a57e-e640664ca683.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.336267] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d40ad28a-d5eb-493e-83f7-9c2de2fc646a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.346142] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 951.346142] env[65107]: value = "task-5103139" [ 951.346142] env[65107]: _type = "Task" [ 951.346142] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.359853] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.364735] env[65107]: DEBUG nova.network.neutron [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Updating instance_info_cache with network_info: [{"id": "853b4d7e-c8ee-4570-a59c-310c21f06168", "address": "fa:16:3e:c7:99:09", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853b4d7e-c8", "ovs_interfaceid": "853b4d7e-c8ee-4570-a59c-310c21f06168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.387359] env[65107]: DEBUG nova.network.neutron [-] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.596910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.596910] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 951.598716] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.860s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.598938] env[65107]: DEBUG nova.objects.instance [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 951.705728] env[65107]: INFO nova.compute.manager [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Took 41.42 seconds to build instance. [ 951.751032] env[65107]: DEBUG oslo_concurrency.lockutils [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] Releasing lock "refresh_cache-d5766a03-054f-40ea-a57e-e640664ca683" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.751301] env[65107]: DEBUG nova.compute.manager [req-0c7ee637-4f99-43ab-b194-9e1338d1b423 req-164464b4-8c9c-46eb-b8c3-6acfb967f0f4 service nova] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Received event network-vif-deleted-860d5e06-e063-471a-8b43-cda8c6bd94b8 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 951.859513] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103139, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.871881] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.871881] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.876503] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "refresh_cache-52bd7ecb-4f53-443e-8916-3ff50dfab2b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.876503] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Instance network_info: |[{"id": "853b4d7e-c8ee-4570-a59c-310c21f06168", "address": "fa:16:3e:c7:99:09", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853b4d7e-c8", "ovs_interfaceid": "853b4d7e-c8ee-4570-a59c-310c21f06168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 951.876503] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:99:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '853b4d7e-c8ee-4570-a59c-310c21f06168', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.883689] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 951.884487] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.884734] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fce03963-bcd2-4cc3-84c6-61d32bafa9cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.906314] env[65107]: INFO nova.compute.manager [-] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Took 1.77 seconds to deallocate network for instance. [ 951.917823] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.917823] env[65107]: value = "task-5103140" [ 951.917823] env[65107]: _type = "Task" [ 951.917823] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.933336] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103140, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.108721] env[65107]: DEBUG nova.compute.utils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 952.117655] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 952.117655] env[65107]: DEBUG nova.network.neutron [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 952.117655] env[65107]: WARNING neutronclient.v2_0.client [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.117655] env[65107]: WARNING neutronclient.v2_0.client [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.117655] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.117655] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.191378] env[65107]: DEBUG nova.policy [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61d0444710414b8eae5e40ed898d8712', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb7fca6907c4c32a1aa6b8023c39afb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 952.208125] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e44cd89-2ccd-48d0-aa94-1df4ff9e2d14 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.943s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.309766] env[65107]: DEBUG nova.compute.manager [req-03ca1b9d-bc59-4018-b7c3-9037ab1c39e6 req-3ef924ce-ae9f-4e31-9932-2146be76da0c service nova] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Received event network-vif-deleted-b58f1ae3-f99e-46a9-81be-e3d4925f3f85 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 952.337367] env[65107]: DEBUG nova.compute.manager [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Received event network-changed-853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 952.337505] env[65107]: DEBUG nova.compute.manager [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Refreshing instance network info cache due to event network-changed-853b4d7e-c8ee-4570-a59c-310c21f06168. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 952.337718] env[65107]: DEBUG oslo_concurrency.lockutils [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Acquiring lock "refresh_cache-52bd7ecb-4f53-443e-8916-3ff50dfab2b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.337914] env[65107]: DEBUG oslo_concurrency.lockutils [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Acquired lock "refresh_cache-52bd7ecb-4f53-443e-8916-3ff50dfab2b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.338131] env[65107]: DEBUG nova.network.neutron [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Refreshing network info cache for port 853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 952.359819] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103139, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762091} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.360162] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] d5766a03-054f-40ea-a57e-e640664ca683/d5766a03-054f-40ea-a57e-e640664ca683.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.360419] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.360648] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53bb1d96-4322-42dd-a6f2-b8f7265f41e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.369652] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 952.369652] env[65107]: value = "task-5103141" [ 952.369652] env[65107]: _type = "Task" [ 952.369652] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.378161] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.378404] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.378614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.379020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.379209] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.383460] env[65107]: INFO nova.compute.manager [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Terminating instance [ 952.385519] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 952.393262] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.419360] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.430863] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103140, 'name': CreateVM_Task, 'duration_secs': 0.436814} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.432621] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.432621] env[65107]: WARNING neutronclient.v2_0.client [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.432797] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.432890] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.433795] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 952.433922] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b11f9af-1832-4add-b370-9f3ea84f265a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.439622] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 952.439622] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b76e89-9360-91e5-e14a-316fe9f276a8" [ 952.439622] env[65107]: _type = "Task" [ 952.439622] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.453506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "3df65e64-0c45-4707-960e-8f1767e2d011" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.453740] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "3df65e64-0c45-4707-960e-8f1767e2d011" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.455134] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b76e89-9360-91e5-e14a-316fe9f276a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.591839] env[65107]: DEBUG nova.network.neutron [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Successfully created port: ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 952.626279] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 952.630311] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1176629e-d081-4a48-b09a-5be5689e0bfb tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.032s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.631536] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.436s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.631767] env[65107]: DEBUG nova.objects.instance [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lazy-loading 'resources' on Instance uuid 4a42392a-a4b4-43b7-a410-0b69441435be {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.842029] env[65107]: WARNING neutronclient.v2_0.client [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.843172] env[65107]: WARNING openstack [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.843758] env[65107]: WARNING openstack [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.880829] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078923} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.881269] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.882127] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9cdc24d-91e7-4707-b25f-539ffb4921f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.900975] env[65107]: DEBUG nova.compute.manager [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 952.901311] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.912213] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] d5766a03-054f-40ea-a57e-e640664ca683/d5766a03-054f-40ea-a57e-e640664ca683.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.915665] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00071c47-7b02-4483-821a-1face1b3b3e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.918492] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61cc2609-6ebe-456e-808b-a1dacc6138a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.938635] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.940018] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf29f350-81d5-4219-995e-34749a209c25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.941934] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 952.941934] env[65107]: value = "task-5103142" [ 952.941934] env[65107]: _type = "Task" [ 952.941934] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.953048] env[65107]: DEBUG oslo_vmware.api [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 952.953048] env[65107]: value = "task-5103143" [ 952.953048] env[65107]: _type = "Task" [ 952.953048] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.959395] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 952.966751] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.970671] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.970826] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Processing image 57f71a90-320c-4d88-8671-5ad259d2a173 {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.971873] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.971873] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.971873] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.971873] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103142, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.972100] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bd4d922-11b1-4861-a44c-e8a2fa603ac2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.979627] env[65107]: DEBUG oslo_vmware.api [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.991922] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.992228] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 952.993093] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7d3d2f5-abb9-4d5a-b380-fe4cf0adabaa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.000241] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 953.000241] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fa9dd7-a73e-4a08-737f-5081c4e79fb9" [ 953.000241] env[65107]: _type = "Task" [ 953.000241] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.009766] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fa9dd7-a73e-4a08-737f-5081c4e79fb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.160038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.160038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.160038] env[65107]: DEBUG nova.compute.manager [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 953.160038] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942b798d-a5d1-4c8f-83e5-c9df07fde614 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.171843] env[65107]: DEBUG nova.compute.manager [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 953.173524] env[65107]: DEBUG nova.objects.instance [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'flavor' on Instance uuid 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.456380] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103142, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.470479] env[65107]: DEBUG oslo_vmware.api [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103143, 'name': PowerOffVM_Task, 'duration_secs': 0.255366} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.470752] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.470913] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.471186] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88600276-cd4c-46e3-bb9b-7c2ca15d764b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.499823] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.517245] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 953.517872] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Fetch image to [datastore1] OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef/OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 953.518095] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Downloading stream optimized image 57f71a90-320c-4d88-8671-5ad259d2a173 to [datastore1] OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef/OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef.vmdk on the data store datastore1 as vApp {{(pid=65107) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 953.518257] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Downloading image file data 57f71a90-320c-4d88-8671-5ad259d2a173 to the ESX as VM named 'OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef' {{(pid=65107) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 953.574603] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.574816] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.575078] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Deleting the datastore file [datastore2] 179e613b-e5a2-4fbc-8fa6-b72769425ff3 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.578156] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f182a7fb-f71e-4bee-8c16-338941d6dcf6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.586327] env[65107]: DEBUG oslo_vmware.api [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for the task: (returnval){ [ 953.586327] env[65107]: value = "task-5103145" [ 953.586327] env[65107]: _type = "Task" [ 953.586327] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.600285] env[65107]: DEBUG oslo_vmware.api [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.628601] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 953.628601] env[65107]: value = "resgroup-9" [ 953.628601] env[65107]: _type = "ResourcePool" [ 953.628601] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 953.631049] env[65107]: WARNING openstack [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.631508] env[65107]: WARNING openstack [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.640279] env[65107]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e37b09c3-5ea2-4ad8-a1a0-e70e807c4fd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.663322] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 953.666293] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cacc7a43-9186-48cc-bfea-cc4495010bf5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.676925] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f5ca81-7071-42a1-b8df-ab847e8299d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.680763] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease: (returnval){ [ 953.680763] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521577df-e150-2bde-0f81-9cbfda259f84" [ 953.680763] env[65107]: _type = "HttpNfcLease" [ 953.680763] env[65107]: } obtained for vApp import into resource pool (val){ [ 953.680763] env[65107]: value = "resgroup-9" [ 953.680763] env[65107]: _type = "ResourcePool" [ 953.680763] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 953.681057] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the lease: (returnval){ [ 953.681057] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521577df-e150-2bde-0f81-9cbfda259f84" [ 953.681057] env[65107]: _type = "HttpNfcLease" [ 953.681057] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.728878] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 953.729218] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 953.729420] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 953.729649] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 953.729817] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 953.730375] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 953.730448] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.730630] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 953.730849] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 953.731105] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 953.731314] env[65107]: DEBUG nova.virt.hardware [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 953.732731] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b35b13-01c8-4a88-975d-93b0693327b4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.736837] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc857493-230c-46cc-9419-fddf0af2392f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.742080] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.742080] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521577df-e150-2bde-0f81-9cbfda259f84" [ 953.742080] env[65107]: _type = "HttpNfcLease" [ 953.742080] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.750423] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5245b9-48a6-4289-ab89-344c296d9fc8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.756267] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2eed81-95e7-440a-bb42-415eb572e486 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.781141] env[65107]: DEBUG nova.compute.provider_tree [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.956385] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103142, 'name': ReconfigVM_Task, 'duration_secs': 0.645978} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.956708] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Reconfigured VM instance instance-0000004e to attach disk [datastore1] d5766a03-054f-40ea-a57e-e640664ca683/d5766a03-054f-40ea-a57e-e640664ca683.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.957448] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffa2e5a1-d63f-4b39-94b3-1c0824c57e0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.965443] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 953.965443] env[65107]: value = "task-5103147" [ 953.965443] env[65107]: _type = "Task" [ 953.965443] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.974885] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103147, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.101068] env[65107]: DEBUG oslo_vmware.api [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Task: {'id': task-5103145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.269076} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.101449] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.101697] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.101924] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.102215] env[65107]: INFO nova.compute.manager [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Took 1.20 seconds to destroy the instance on the hypervisor. [ 954.102511] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 954.102887] env[65107]: DEBUG nova.compute.manager [-] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 954.102887] env[65107]: DEBUG nova.network.neutron [-] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 954.103703] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.107325] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.107637] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.197811] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.198829] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.198829] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521577df-e150-2bde-0f81-9cbfda259f84" [ 954.198829] env[65107]: _type = "HttpNfcLease" [ 954.198829] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 954.199201] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a1115f7-cca3-4d3b-82fb-314c4528f176 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.210105] env[65107]: DEBUG oslo_vmware.api [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 954.210105] env[65107]: value = "task-5103148" [ 954.210105] env[65107]: _type = "Task" [ 954.210105] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.222984] env[65107]: DEBUG oslo_vmware.api [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.284248] env[65107]: DEBUG nova.scheduler.client.report [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.346186] env[65107]: DEBUG nova.network.neutron [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Successfully updated port: ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 954.447604] env[65107]: WARNING neutronclient.v2_0.client [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.448367] env[65107]: WARNING openstack [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.449641] env[65107]: WARNING openstack [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.458763] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.478770] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103147, 'name': Rename_Task, 'duration_secs': 0.316919} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.479348] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.479737] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40df2423-741c-4cb7-a97a-5abeb0d9c053 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.488054] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 954.488054] env[65107]: value = "task-5103149" [ 954.488054] env[65107]: _type = "Task" [ 954.488054] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.498844] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.700956] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.700956] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521577df-e150-2bde-0f81-9cbfda259f84" [ 954.700956] env[65107]: _type = "HttpNfcLease" [ 954.700956] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.700956] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.700956] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521577df-e150-2bde-0f81-9cbfda259f84" [ 954.700956] env[65107]: _type = "HttpNfcLease" [ 954.700956] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 954.700956] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0840ba6b-2298-4e69-a435-1ad6a1ef9664 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.710586] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528dcb3c-5e34-efc4-e3c5-e41daf160d6b/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.711485] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528dcb3c-5e34-efc4-e3c5-e41daf160d6b/disk-0.vmdk. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 954.790855] env[65107]: DEBUG oslo_vmware.api [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103148, 'name': PowerOffVM_Task, 'duration_secs': 0.362148} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.792428] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.792689] env[65107]: DEBUG nova.compute.manager [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 954.799196] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeef14a0-0394-4211-9d5d-87318719856b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.802404] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.171s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.807154] env[65107]: DEBUG nova.network.neutron [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Updated VIF entry in instance network info cache for port 853b4d7e-c8ee-4570-a59c-310c21f06168. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 954.807401] env[65107]: DEBUG nova.network.neutron [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Updating instance_info_cache with network_info: [{"id": "853b4d7e-c8ee-4570-a59c-310c21f06168", "address": "fa:16:3e:c7:99:09", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853b4d7e-c8", "ovs_interfaceid": "853b4d7e-c8ee-4570-a59c-310c21f06168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 954.812043] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4ea8ff8c-1659-4240-8cc1-84fa38269354 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.814751] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.202s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.818692] env[65107]: INFO nova.compute.claims [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.851089] env[65107]: INFO nova.scheduler.client.report [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Deleted allocations for instance 4a42392a-a4b4-43b7-a410-0b69441435be [ 954.852201] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.852362] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.852531] env[65107]: DEBUG nova.network.neutron [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 955.008990] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103149, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.225617] env[65107]: DEBUG nova.network.neutron [-] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.290901] env[65107]: DEBUG nova.compute.manager [req-488441c6-1c86-4a79-a6a2-abe48f717930 req-afc81ee1-f633-4f3d-b7e2-baf514303a99 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received event network-vif-plugged-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 955.291334] env[65107]: DEBUG oslo_concurrency.lockutils [req-488441c6-1c86-4a79-a6a2-abe48f717930 req-afc81ee1-f633-4f3d-b7e2-baf514303a99 service nova] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.291596] env[65107]: DEBUG oslo_concurrency.lockutils [req-488441c6-1c86-4a79-a6a2-abe48f717930 req-afc81ee1-f633-4f3d-b7e2-baf514303a99 service nova] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.291888] env[65107]: DEBUG oslo_concurrency.lockutils [req-488441c6-1c86-4a79-a6a2-abe48f717930 req-afc81ee1-f633-4f3d-b7e2-baf514303a99 service nova] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.291980] env[65107]: DEBUG nova.compute.manager [req-488441c6-1c86-4a79-a6a2-abe48f717930 req-afc81ee1-f633-4f3d-b7e2-baf514303a99 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] No waiting events found dispatching network-vif-plugged-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 955.292160] env[65107]: WARNING nova.compute.manager [req-488441c6-1c86-4a79-a6a2-abe48f717930 req-afc81ee1-f633-4f3d-b7e2-baf514303a99 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received unexpected event network-vif-plugged-ed79bc21-5c5f-44e5-ba58-55e43c814088 for instance with vm_state building and task_state spawning. [ 955.322447] env[65107]: DEBUG oslo_concurrency.lockutils [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] Releasing lock "refresh_cache-52bd7ecb-4f53-443e-8916-3ff50dfab2b6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.322737] env[65107]: DEBUG nova.compute.manager [req-3b421fb8-6e53-403f-8c7b-8d0878f4a402 req-0c56cbe4-f4fa-4cdd-a1a5-ee7247a3b948 service nova] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Received event network-vif-deleted-875128de-2b86-4f3c-bef3-92733849587c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 955.336220] env[65107]: DEBUG oslo_concurrency.lockutils [None req-717d8943-df1d-4e31-bf1c-f9ffdcb01b6c tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.178s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.364871] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.366059] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.382132] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be29e098-2a5e-4004-8d70-cad694ca5dc2 tempest-ServersListShow296Test-1456045782 tempest-ServersListShow296Test-1456045782-project-member] Lock "4a42392a-a4b4-43b7-a410-0b69441435be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.492s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.455762] env[65107]: DEBUG nova.network.neutron [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 955.500023] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.500023] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.512239] env[65107]: DEBUG oslo_vmware.api [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103149, 'name': PowerOnVM_Task, 'duration_secs': 0.701151} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.514362] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.514578] env[65107]: INFO nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Took 10.02 seconds to spawn the instance on the hypervisor. [ 955.514754] env[65107]: DEBUG nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 955.515684] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee97433-1078-442b-8673-71a68c240e80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.576067] env[65107]: WARNING neutronclient.v2_0.client [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.576837] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.577217] env[65107]: WARNING openstack [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.606080] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 955.606080] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528dcb3c-5e34-efc4-e3c5-e41daf160d6b/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 955.607380] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6713829-73d5-4aa9-b3ec-aae4b5cce68c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.617699] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528dcb3c-5e34-efc4-e3c5-e41daf160d6b/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 955.617699] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528dcb3c-5e34-efc4-e3c5-e41daf160d6b/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 955.617699] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b2722e1a-9bde-4ec2-9ac4-ad1d6ce48804 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.678196] env[65107]: DEBUG nova.network.neutron [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.730597] env[65107]: INFO nova.compute.manager [-] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Took 1.63 seconds to deallocate network for instance. [ 955.804542] env[65107]: DEBUG oslo_vmware.rw_handles [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528dcb3c-5e34-efc4-e3c5-e41daf160d6b/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 955.805439] env[65107]: INFO nova.virt.vmwareapi.images [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Downloaded image file data 57f71a90-320c-4d88-8671-5ad259d2a173 [ 955.807332] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4895be50-df09-43e2-a57e-269ba1b07493 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.837550] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d60d0836-0b3f-4652-a62b-43a66e323e4d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.864199] env[65107]: INFO nova.virt.vmwareapi.images [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] The imported VM was unregistered [ 955.869022] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 955.869333] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating directory with path [datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.873068] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-752f7ff2-838b-429f-bea8-bf390cfcaa2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.894381] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created directory with path [datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.894663] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef/OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef.vmdk to [datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk. {{(pid=65107) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 955.894943] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f73b979c-f225-43f3-9ed3-95010337c306 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.903857] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 955.903857] env[65107]: value = "task-5103151" [ 955.903857] env[65107]: _type = "Task" [ 955.903857] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.915434] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.046667] env[65107]: INFO nova.compute.manager [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Took 38.04 seconds to build instance. [ 956.184022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 956.184022] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Instance network_info: |[{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 956.184022] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:25:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '27138a4c-60c9-45fb-bf37-4c2f765315a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed79bc21-5c5f-44e5-ba58-55e43c814088', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.190413] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 956.194212] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.194474] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.194696] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.195454] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.195653] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.197205] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.197917] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18971a10-a123-408f-afdd-37804b569865 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.213404] env[65107]: INFO nova.compute.manager [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Terminating instance [ 956.225015] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.225015] env[65107]: value = "task-5103152" [ 956.225015] env[65107]: _type = "Task" [ 956.225015] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.240889] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.242157] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103152, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.354346] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550b77af-82bc-418e-a2c3-555647ebf7e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.367588] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9712134-a219-4e78-b370-24b8e0ccd03b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.409874] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0680d14f-8493-41a4-ae8f-172d36bddeb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.422312] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6ad6bf-a92a-4c62-a22d-60d987d7e284 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.427029] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.440576] env[65107]: DEBUG nova.compute.provider_tree [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.551784] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4d723e27-f208-4a02-bb4c-31bf80953c99 tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "d5766a03-054f-40ea-a57e-e640664ca683" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.555s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.726091] env[65107]: DEBUG nova.compute.manager [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 956.726091] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.727460] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa6875b-5208-4133-adb1-4cf80ed389b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.754806] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103152, 'name': CreateVM_Task, 'duration_secs': 0.416154} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.757273] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.757629] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.758247] env[65107]: WARNING neutronclient.v2_0.client [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.758627] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.758776] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 956.759115] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 956.759368] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcb2c324-523a-4e02-91ef-6b55e06e14a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.761532] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1259c1b-6ace-4578-ba16-2bb162318108 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.768869] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 956.768869] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b846dc-f40e-1e34-9f60-d08c2f369ff3" [ 956.768869] env[65107]: _type = "Task" [ 956.768869] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.783123] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b846dc-f40e-1e34-9f60-d08c2f369ff3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.908934] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.908934] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.910100] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.910353] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.910641] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleting the datastore file [datastore1] 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.914776] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27318545-0bbb-4384-85f2-a3d8a8660cb9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.927288] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.929421] env[65107]: DEBUG oslo_vmware.api [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 956.929421] env[65107]: value = "task-5103154" [ 956.929421] env[65107]: _type = "Task" [ 956.929421] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.939536] env[65107]: DEBUG oslo_vmware.api [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.944736] env[65107]: DEBUG nova.scheduler.client.report [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 957.227670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "d5766a03-054f-40ea-a57e-e640664ca683" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.227670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "d5766a03-054f-40ea-a57e-e640664ca683" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.227670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "d5766a03-054f-40ea-a57e-e640664ca683-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.227964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "d5766a03-054f-40ea-a57e-e640664ca683-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.227964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "d5766a03-054f-40ea-a57e-e640664ca683-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.230456] env[65107]: INFO nova.compute.manager [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Terminating instance [ 957.282552] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b846dc-f40e-1e34-9f60-d08c2f369ff3, 'name': SearchDatastore_Task, 'duration_secs': 0.046362} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.282970] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 957.283295] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.283595] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.283763] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.283996] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.284348] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3ec9e4f-9751-4f31-a1f1-452e6f9cccd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.298800] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.299037] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.299841] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caa7da47-2a37-4da5-8a2e-04b75c1198f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.308196] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 957.308196] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de3326-fa2f-1ed3-6462-58502309d4d3" [ 957.308196] env[65107]: _type = "Task" [ 957.308196] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.320948] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de3326-fa2f-1ed3-6462-58502309d4d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.338342] env[65107]: DEBUG nova.compute.manager [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 957.338480] env[65107]: DEBUG nova.compute.manager [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing instance network info cache due to event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 957.338940] env[65107]: DEBUG oslo_concurrency.lockutils [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.339146] env[65107]: DEBUG oslo_concurrency.lockutils [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.339309] env[65107]: DEBUG nova.network.neutron [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 957.413805] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 957.422683] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.440667] env[65107]: DEBUG oslo_vmware.api [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.449774] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.450334] env[65107]: DEBUG nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 957.453024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.459s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.454473] env[65107]: INFO nova.compute.claims [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.735000] env[65107]: DEBUG nova.compute.manager [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 957.735401] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.736409] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f167d8e-7411-4026-9924-4e095cf20d80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.747067] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.747067] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46cf63a1-f8d7-46c3-888f-45c20a99e4b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.755202] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 957.755202] env[65107]: value = "task-5103155" [ 957.755202] env[65107]: _type = "Task" [ 957.755202] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.767252] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.821153] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de3326-fa2f-1ed3-6462-58502309d4d3, 'name': SearchDatastore_Task, 'duration_secs': 0.046658} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.822438] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fe46c79-0cb6-49d8-b355-776f19c98111 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.831728] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 957.831728] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523fe7fb-93e2-10b2-3982-a2610e129738" [ 957.831728] env[65107]: _type = "Task" [ 957.831728] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.841766] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523fe7fb-93e2-10b2-3982-a2610e129738, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.845690] env[65107]: WARNING neutronclient.v2_0.client [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.846650] env[65107]: WARNING openstack [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.847124] env[65107]: WARNING openstack [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.923937] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.945987] env[65107]: DEBUG oslo_vmware.api [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.948159] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.963204] env[65107]: DEBUG nova.compute.utils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 957.965635] env[65107]: DEBUG nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 957.975173] env[65107]: WARNING openstack [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.975691] env[65107]: WARNING openstack [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.053970] env[65107]: WARNING neutronclient.v2_0.client [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.054711] env[65107]: WARNING openstack [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.055075] env[65107]: WARNING openstack [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.175840] env[65107]: DEBUG nova.network.neutron [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updated VIF entry in instance network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 958.176276] env[65107]: DEBUG nova.network.neutron [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 958.272199] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103155, 'name': PowerOffVM_Task, 'duration_secs': 0.222805} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.272525] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.272749] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.273093] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07f7faf1-3f10-4bd2-a7f7-1153d9270213 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.342080] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.342511] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.342511] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Deleting the datastore file [datastore1] d5766a03-054f-40ea-a57e-e640664ca683 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.346682] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df8fcd46-4abd-4421-8669-d9454ce222bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.348798] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523fe7fb-93e2-10b2-3982-a2610e129738, 'name': SearchDatastore_Task, 'duration_secs': 0.048969} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.349096] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.349361] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e12cb11f-7d6f-47d6-a31e-bf2654174d38.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 958.350045] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2770ce68-6126-45b2-81ab-7478c49d4678 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.357339] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for the task: (returnval){ [ 958.357339] env[65107]: value = "task-5103157" [ 958.357339] env[65107]: _type = "Task" [ 958.357339] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.362718] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 958.362718] env[65107]: value = "task-5103158" [ 958.362718] env[65107]: _type = "Task" [ 958.362718] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.370464] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.374134] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103158, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.421025] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.444509] env[65107]: DEBUG oslo_vmware.api [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.468998] env[65107]: DEBUG nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 958.682181] env[65107]: DEBUG oslo_concurrency.lockutils [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.682458] env[65107]: DEBUG nova.compute.manager [req-5dd78cd7-d5d6-4620-a129-a4e7e1d0afc8 req-5b7d16d5-ab3e-4f02-b0bc-00b583854f47 service nova] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Received event network-vif-deleted-bec67e7b-de92-407d-aef5-04d4430d3ee9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 958.872052] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.877420] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103158, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.892402] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133511dd-a3ff-4d7e-9ba4-faf053512bcd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.900330] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbcbfb8-546f-49fd-91ec-98aa252ca3af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.942764] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac4a341-582f-4810-a1e0-b18458e64c4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.954857] env[65107]: DEBUG oslo_vmware.api [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103154, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.752471} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.956639] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103151, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.896034} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.957516] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.957741] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.957928] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.958120] env[65107]: INFO nova.compute.manager [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Took 2.23 seconds to destroy the instance on the hypervisor. [ 958.958369] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 958.958577] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef/OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef.vmdk to [datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk. [ 958.958746] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Cleaning up location [datastore1] OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 958.958902] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_119458b5-890d-4d27-b1da-cdbc47a7c5ef {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.959216] env[65107]: DEBUG nova.compute.manager [-] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 958.959315] env[65107]: DEBUG nova.network.neutron [-] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 958.959565] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.960121] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.960381] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.967165] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78463f47-7fad-4b1e-99c3-0f11512f4656 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.970348] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c734c4-c465-4b1b-abe8-cec24608a784 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.991276] env[65107]: DEBUG nova.compute.provider_tree [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.995182] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 958.995182] env[65107]: value = "task-5103159" [ 958.995182] env[65107]: _type = "Task" [ 958.995182] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.002394] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.045286] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.373220] env[65107]: DEBUG oslo_vmware.api [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Task: {'id': task-5103157, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.774218} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.374136] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.375145] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.375458] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.376073] env[65107]: INFO nova.compute.manager [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Took 1.64 seconds to destroy the instance on the hypervisor. [ 959.376073] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 959.376378] env[65107]: DEBUG nova.compute.manager [-] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 959.376517] env[65107]: DEBUG nova.network.neutron [-] [instance: d5766a03-054f-40ea-a57e-e640664ca683] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 959.376875] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.377476] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.377739] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.390726] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103158, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.392138] env[65107]: DEBUG nova.compute.manager [req-86a83713-8bd8-4cfd-b168-139885dc0b49 req-19e340c0-0305-466c-a6a4-a2555896e46f service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Received event network-vif-deleted-1c2c4d44-fde7-4882-9b77-f18455fd3786 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 959.392365] env[65107]: INFO nova.compute.manager [req-86a83713-8bd8-4cfd-b168-139885dc0b49 req-19e340c0-0305-466c-a6a4-a2555896e46f service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Neutron deleted interface 1c2c4d44-fde7-4882-9b77-f18455fd3786; detaching it from the instance and deleting it from the info cache [ 959.392595] env[65107]: DEBUG nova.network.neutron [req-86a83713-8bd8-4cfd-b168-139885dc0b49 req-19e340c0-0305-466c-a6a4-a2555896e46f service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.423069] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.479878] env[65107]: DEBUG nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 959.500752] env[65107]: DEBUG nova.scheduler.client.report [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.510499] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1163} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.512757] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 959.513137] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 959.513137] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 959.513310] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 959.513521] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 959.514765] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 959.514765] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.514765] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 959.514765] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 959.514765] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 959.514765] env[65107]: DEBUG nova.virt.hardware [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 959.514765] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.515191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.515191] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk to [datastore1] 52bd7ecb-4f53-443e-8916-3ff50dfab2b6/52bd7ecb-4f53-443e-8916-3ff50dfab2b6.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.516647] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b212ccb-c1e1-41aa-bf5e-dacb71a4e153 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.519240] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3db74d1-e5cc-410b-a8ff-6f804ffa4968 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.529044] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c85e65-9af9-4cf3-b133-f05b0d425570 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.533546] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 959.533546] env[65107]: value = "task-5103160" [ 959.533546] env[65107]: _type = "Task" [ 959.533546] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.547965] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.554073] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Creating folder: Project (a347dd83e3b24c41b83bb24196acfad9). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 959.555134] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d8eaf980-dc0a-4da8-988b-01dae768f0f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.562396] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.572177] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Created folder: Project (a347dd83e3b24c41b83bb24196acfad9) in parent group-v992574. [ 959.572446] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Creating folder: Instances. Parent ref: group-v992804. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 959.572798] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f266f5cd-a406-4cf2-80f0-ed1d26aecf9f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.584034] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Created folder: Instances in parent group-v992804. [ 959.584260] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 959.588018] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.588018] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cd855aa-25c4-4bfc-8651-4dbbe983406e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.615945] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.615945] env[65107]: value = "task-5103163" [ 959.615945] env[65107]: _type = "Task" [ 959.615945] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.625969] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103163, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.635466] env[65107]: DEBUG nova.compute.manager [req-b4299e5f-5828-471d-950e-cbf6a52f9e34 req-11b47792-00f5-4a7d-8ee7-aaf66e7ddaa4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Received event network-vif-deleted-c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 959.635686] env[65107]: INFO nova.compute.manager [req-b4299e5f-5828-471d-950e-cbf6a52f9e34 req-11b47792-00f5-4a7d-8ee7-aaf66e7ddaa4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Neutron deleted interface c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6; detaching it from the instance and deleting it from the info cache [ 959.635861] env[65107]: DEBUG nova.network.neutron [req-b4299e5f-5828-471d-950e-cbf6a52f9e34 req-11b47792-00f5-4a7d-8ee7-aaf66e7ddaa4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.803087] env[65107]: DEBUG nova.network.neutron [-] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.880170] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103158, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.111018} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.880391] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e12cb11f-7d6f-47d6-a31e-bf2654174d38.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 959.880622] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 959.880895] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce2f8f5b-a71f-4c89-bd21-2938eee2c7fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.888705] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 959.888705] env[65107]: value = "task-5103164" [ 959.888705] env[65107]: _type = "Task" [ 959.888705] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.899458] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103164, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.899824] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c014e4b1-55ab-4ba5-9858-2f215674748b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.912459] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e332a62-3e68-45cd-bbe5-6995dd97ffa3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.952659] env[65107]: DEBUG nova.compute.manager [req-86a83713-8bd8-4cfd-b168-139885dc0b49 req-19e340c0-0305-466c-a6a4-a2555896e46f service nova] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Detach interface failed, port_id=1c2c4d44-fde7-4882-9b77-f18455fd3786, reason: Instance 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 960.006375] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.006953] env[65107]: DEBUG nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 960.010657] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.274s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.010888] env[65107]: DEBUG nova.objects.instance [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lazy-loading 'resources' on Instance uuid 70172f4f-2658-4bc1-857f-0dffde2d5ab0 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.049239] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.118279] env[65107]: DEBUG nova.network.neutron [-] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 960.130999] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103163, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.140623] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d44b7ed3-bc9f-41d3-8b1f-e8131ef67cfd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.153774] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edc93c5-29f2-4953-84a1-9753134d5bb6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.190804] env[65107]: DEBUG nova.compute.manager [req-b4299e5f-5828-471d-950e-cbf6a52f9e34 req-11b47792-00f5-4a7d-8ee7-aaf66e7ddaa4 service nova] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Detach interface failed, port_id=c8e399f8-2c17-4218-97fe-e9f1a1ad2eb6, reason: Instance d5766a03-054f-40ea-a57e-e640664ca683 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 960.306054] env[65107]: INFO nova.compute.manager [-] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Took 1.35 seconds to deallocate network for instance. [ 960.402522] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081209} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.402522] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.403315] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574e484b-90a2-4eab-a3e3-4907395b584c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.431585] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e12cb11f-7d6f-47d6-a31e-bf2654174d38.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.431992] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4a177b0-a771-4cb6-a2f3-a912ee662fd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.457269] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 960.457269] env[65107]: value = "task-5103165" [ 960.457269] env[65107]: _type = "Task" [ 960.457269] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.469402] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103165, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.514931] env[65107]: DEBUG nova.compute.utils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 960.516646] env[65107]: DEBUG nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 960.550133] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.626438] env[65107]: INFO nova.compute.manager [-] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Took 1.25 seconds to deallocate network for instance. [ 960.637378] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103163, 'name': CreateVM_Task, 'duration_secs': 0.545042} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.637562] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.638103] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.638265] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.638595] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 960.638930] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb64b3c8-3528-448c-a06f-ed386097ebc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.651071] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 960.651071] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52efbe92-124c-d232-3a0e-865a41640b35" [ 960.651071] env[65107]: _type = "Task" [ 960.651071] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.663559] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52efbe92-124c-d232-3a0e-865a41640b35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.814458] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 960.926466] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5fd1cb-34a7-47ca-a4f8-99101c7c38d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.937339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753e224f-41ba-4841-95c7-99935f971634 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.978483] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659f6103-c9ec-4b9c-b736-e4653ac1d762 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.992115] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feecf4ae-ab77-4799-ab05-f83aad93c29e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.996709] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103165, 'name': ReconfigVM_Task, 'duration_secs': 0.525046} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.996991] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfigured VM instance instance-00000050 to attach disk [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e12cb11f-7d6f-47d6-a31e-bf2654174d38.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.998202] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05a7e05a-694f-4d4c-a4bd-41903a3f4987 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.012101] env[65107]: DEBUG nova.compute.provider_tree [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.014406] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 961.014406] env[65107]: value = "task-5103166" [ 961.014406] env[65107]: _type = "Task" [ 961.014406] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.018670] env[65107]: DEBUG nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 961.029710] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103166, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.047803] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.135059] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.167063] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52efbe92-124c-d232-3a0e-865a41640b35, 'name': SearchDatastore_Task, 'duration_secs': 0.050081} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.167063] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 961.167319] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.167598] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.167781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.168035] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.168356] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba8560f6-5525-4be6-a3de-05f420423b29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.183884] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.184072] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.185024] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c2f8528-b0f0-4ced-bada-61604b83ba48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.191168] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 961.191168] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a058aa-2d0a-db2f-0ebe-51c3f92ce424" [ 961.191168] env[65107]: _type = "Task" [ 961.191168] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.200350] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a058aa-2d0a-db2f-0ebe-51c3f92ce424, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.516836] env[65107]: DEBUG nova.scheduler.client.report [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 961.537548] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103166, 'name': Rename_Task, 'duration_secs': 0.170983} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.538892] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.543056] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6439bc3e-1e59-4c5a-9e03-e732c39f51f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.552893] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.555199] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 961.555199] env[65107]: value = "task-5103167" [ 961.555199] env[65107]: _type = "Task" [ 961.555199] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.568691] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103167, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.708040] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a058aa-2d0a-db2f-0ebe-51c3f92ce424, 'name': SearchDatastore_Task, 'duration_secs': 0.046791} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.708040] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bb91733-bd56-4584-890c-d38e97800c50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.718013] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 961.718013] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5224df77-6e60-20dd-c4fd-b51baacf8fc6" [ 961.718013] env[65107]: _type = "Task" [ 961.718013] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.731334] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5224df77-6e60-20dd-c4fd-b51baacf8fc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.026597] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.016s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 962.029326] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.562s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 962.030979] env[65107]: INFO nova.compute.claims [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.034958] env[65107]: DEBUG nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 962.051205] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.056022] env[65107]: INFO nova.scheduler.client.report [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Deleted allocations for instance 70172f4f-2658-4bc1-857f-0dffde2d5ab0 [ 962.075474] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103167, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.078765] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 962.078765] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 962.079070] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 962.079614] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 962.079614] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 962.079614] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 962.079769] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.079920] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 962.080134] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 962.081156] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 962.081156] env[65107]: DEBUG nova.virt.hardware [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 962.081466] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a509f68-d582-4204-a4a5-8fd2b2579ce8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.092325] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6de906-7c98-4d0a-bd71-6d9d582662a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.115549] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.123057] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 962.123330] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.123637] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c98a699e-b967-4912-b52f-35710025b28c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.145840] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.145840] env[65107]: value = "task-5103168" [ 962.145840] env[65107]: _type = "Task" [ 962.145840] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.156153] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103168, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.231476] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5224df77-6e60-20dd-c4fd-b51baacf8fc6, 'name': SearchDatastore_Task, 'duration_secs': 0.053347} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.231476] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.231476] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] c5fb4b1c-5540-4645-b996-f397729beb23/c5fb4b1c-5540-4645-b996-f397729beb23.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.231476] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4826dbc-7ac2-4ef5-80bd-fc60566472ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.238725] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 962.238725] env[65107]: value = "task-5103169" [ 962.238725] env[65107]: _type = "Task" [ 962.238725] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.249017] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103169, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.554678] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103160, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.536803} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.555021] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/57f71a90-320c-4d88-8671-5ad259d2a173/57f71a90-320c-4d88-8671-5ad259d2a173.vmdk to [datastore1] 52bd7ecb-4f53-443e-8916-3ff50dfab2b6/52bd7ecb-4f53-443e-8916-3ff50dfab2b6.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 962.555920] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86cfea6-2c03-4bec-8b5c-d90af19641b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.571079] env[65107]: DEBUG oslo_vmware.api [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103167, 'name': PowerOnVM_Task, 'duration_secs': 0.60963} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.571653] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e9028fbe-049e-4254-bcac-c099b5c79187 tempest-ServersNegativeTestMultiTenantJSON-1019907313 tempest-ServersNegativeTestMultiTenantJSON-1019907313-project-member] Lock "70172f4f-2658-4bc1-857f-0dffde2d5ab0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.392s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 962.582961] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.582961] env[65107]: INFO nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Took 8.92 seconds to spawn the instance on the hypervisor. [ 962.583195] env[65107]: DEBUG nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 962.593107] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 52bd7ecb-4f53-443e-8916-3ff50dfab2b6/52bd7ecb-4f53-443e-8916-3ff50dfab2b6.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.595116] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcb7c00-28e8-4efc-a490-9f16d672acd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.598013] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fc83242-af59-4310-8002-5accc5d56479 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.618622] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 962.618622] env[65107]: value = "task-5103170" [ 962.618622] env[65107]: _type = "Task" [ 962.618622] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.635599] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103170, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.658528] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103168, 'name': CreateVM_Task, 'duration_secs': 0.362527} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.658696] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.659198] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.659379] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.659714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 962.659984] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe79cad5-d8c7-4521-8ccc-f08c93ddc49d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.666347] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 962.666347] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dd3193-8d5b-109e-e31e-1fd261ff14b1" [ 962.666347] env[65107]: _type = "Task" [ 962.666347] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.678248] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dd3193-8d5b-109e-e31e-1fd261ff14b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.751121] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103169, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.135028] env[65107]: INFO nova.compute.manager [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Took 33.87 seconds to build instance. [ 963.139269] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103170, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.179638] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dd3193-8d5b-109e-e31e-1fd261ff14b1, 'name': SearchDatastore_Task, 'duration_secs': 0.05592} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.179990] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.180245] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.180484] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.180623] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 963.180805] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.181098] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cec47d6a-fe47-4340-80f7-3aea07c4bdb5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.194707] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.194707] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.195567] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f44cc803-d5f3-418f-af8c-66e26aff2457 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.201904] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 963.201904] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52985b71-f37f-688f-9f60-220daa390437" [ 963.201904] env[65107]: _type = "Task" [ 963.201904] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.213675] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52985b71-f37f-688f-9f60-220daa390437, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.250673] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103169, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534106} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.253584] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] c5fb4b1c-5540-4645-b996-f397729beb23/c5fb4b1c-5540-4645-b996-f397729beb23.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.253854] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.254542] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef141d40-b1e0-4d86-a49c-237b1a5b94b6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.261727] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 963.261727] env[65107]: value = "task-5103171" [ 963.261727] env[65107]: _type = "Task" [ 963.261727] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.274554] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.474291] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233e8c10-d896-4dbf-8a28-dba18c6ccf44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.483728] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c123ab5-6634-457c-953d-bed1b8b4b457 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.517135] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558de7ce-61fd-4a71-9a22-78a82d7653ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.524614] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82072010-0a63-4073-937a-5328dc1fa619 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.540688] env[65107]: DEBUG nova.compute.provider_tree [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.632635] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103170, 'name': ReconfigVM_Task, 'duration_secs': 0.959897} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.632989] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 52bd7ecb-4f53-443e-8916-3ff50dfab2b6/52bd7ecb-4f53-443e-8916-3ff50dfab2b6.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.633634] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24f55e7d-8fc9-401d-bf0f-93319c1a207b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.640151] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33884b0f-bec1-4c06-b6b7-8723676e75dd tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.380s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.641820] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 963.641820] env[65107]: value = "task-5103172" [ 963.641820] env[65107]: _type = "Task" [ 963.641820] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.651015] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103172, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.714031] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52985b71-f37f-688f-9f60-220daa390437, 'name': SearchDatastore_Task, 'duration_secs': 0.011118} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.714722] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07d58ddb-b9b8-4dcb-8df2-4f407ad5ed49 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.721018] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 963.721018] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5296a575-1d92-4cee-7fbe-68b2be8c969e" [ 963.721018] env[65107]: _type = "Task" [ 963.721018] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.731469] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5296a575-1d92-4cee-7fbe-68b2be8c969e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.777187] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103171, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100841} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.777810] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.780345] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41f6e8e-84c0-4336-8f6f-dd41defef472 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.804764] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] c5fb4b1c-5540-4645-b996-f397729beb23/c5fb4b1c-5540-4645-b996-f397729beb23.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.805574] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1af70f9-4b00-4479-a202-d008eb9c3f98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.829101] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 963.829101] env[65107]: value = "task-5103173" [ 963.829101] env[65107]: _type = "Task" [ 963.829101] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.840084] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103173, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.047030] env[65107]: DEBUG nova.scheduler.client.report [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 964.153057] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103172, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.233752] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5296a575-1d92-4cee-7fbe-68b2be8c969e, 'name': SearchDatastore_Task, 'duration_secs': 0.010816} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.235293] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 964.235293] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 964.235293] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-262ee82c-34b6-4220-b471-c8623fcfcdf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.242763] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 964.242763] env[65107]: value = "task-5103174" [ 964.242763] env[65107]: _type = "Task" [ 964.242763] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.251868] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103174, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.344714] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103173, 'name': ReconfigVM_Task, 'duration_secs': 0.289292} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.344876] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Reconfigured VM instance instance-00000051 to attach disk [datastore2] c5fb4b1c-5540-4645-b996-f397729beb23/c5fb4b1c-5540-4645-b996-f397729beb23.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.345503] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-666ffbc1-b80a-4543-b1f0-e33b2aed3c93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.353712] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 964.353712] env[65107]: value = "task-5103175" [ 964.353712] env[65107]: _type = "Task" [ 964.353712] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.366110] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103175, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.553136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.553740] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 964.556584] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.826s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.556815] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.559043] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.088s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.559286] env[65107]: DEBUG nova.objects.instance [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lazy-loading 'resources' on Instance uuid cddb14db-34c3-4783-b21e-e0b4623a3a9e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.607362] env[65107]: INFO nova.scheduler.client.report [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted allocations for instance 71e6e280-50c3-4655-a6a3-40ea4c650d96 [ 964.659096] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103172, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.759462] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103174, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512769} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.759758] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.760222] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.760573] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a93ae46-ecfb-4678-9da4-31dc7b6382a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.768878] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 964.768878] env[65107]: value = "task-5103176" [ 964.768878] env[65107]: _type = "Task" [ 964.768878] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.824080] env[65107]: DEBUG nova.compute.manager [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 964.824080] env[65107]: DEBUG nova.compute.manager [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing instance network info cache due to event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 964.824080] env[65107]: DEBUG oslo_concurrency.lockutils [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.824080] env[65107]: DEBUG oslo_concurrency.lockutils [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.824080] env[65107]: DEBUG nova.network.neutron [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 964.867376] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103175, 'name': Rename_Task, 'duration_secs': 0.427345} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.867699] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 964.868035] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3dcad6c-f556-49a0-95ce-ec5fb63db54d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.878300] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 964.878300] env[65107]: value = "task-5103177" [ 964.878300] env[65107]: _type = "Task" [ 964.878300] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.888927] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.065741] env[65107]: DEBUG nova.compute.utils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 965.071590] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 965.071799] env[65107]: DEBUG nova.network.neutron [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 965.072164] env[65107]: WARNING neutronclient.v2_0.client [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.072521] env[65107]: WARNING neutronclient.v2_0.client [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.073638] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.073638] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.125720] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb6b1386-b2da-4655-9886-95af158110b3 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "71e6e280-50c3-4655-a6a3-40ea4c650d96" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.202s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.160846] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103172, 'name': Rename_Task, 'duration_secs': 1.438592} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.161958] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.162280] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5d19fa0-76e8-4e87-b376-f8210d85b26f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.170680] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 965.170680] env[65107]: value = "task-5103178" [ 965.170680] env[65107]: _type = "Task" [ 965.170680] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.188625] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.216740] env[65107]: DEBUG nova.policy [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd104cf9ef26442a095215b61f39852b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '746d58fc48e64fc58ac57ad38b8051f3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 965.283718] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078697} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.284043] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.284874] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2eebc0-3cc7-4176-ac80-0bf08e57f95b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.313847] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.317734] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-613a564a-aade-40a0-87cf-c0434b97d515 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.338275] env[65107]: WARNING neutronclient.v2_0.client [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.339077] env[65107]: WARNING openstack [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.340046] env[65107]: WARNING openstack [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.359252] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 965.359252] env[65107]: value = "task-5103179" [ 965.359252] env[65107]: _type = "Task" [ 965.359252] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.367309] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103179, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.392454] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103177, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.575542] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 965.586036] env[65107]: DEBUG nova.network.neutron [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Successfully created port: 69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 965.613280] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc1c353-81e7-446e-b092-702b841f7431 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.624846] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfc4e3a-a546-4586-9530-1b17631134ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.662413] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cdbf46-e3f5-4314-b7dd-0d69c4a05de2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.677073] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fd4817-c3de-4294-bab8-c543f2194056 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.689024] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103178, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.697089] env[65107]: DEBUG nova.compute.provider_tree [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.867909] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103179, 'name': ReconfigVM_Task, 'duration_secs': 0.330194} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.868346] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.869021] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fc3ef4a-a3b1-4b2e-9050-560811579a97 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.878081] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 965.878081] env[65107]: value = "task-5103180" [ 965.878081] env[65107]: _type = "Task" [ 965.878081] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.892969] env[65107]: DEBUG oslo_vmware.api [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103177, 'name': PowerOnVM_Task, 'duration_secs': 0.650143} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.896729] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 965.896998] env[65107]: INFO nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Took 6.42 seconds to spawn the instance on the hypervisor. [ 965.897282] env[65107]: DEBUG nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 965.897478] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103180, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.898255] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c469c8c-a784-4d1b-88df-ee329af5b52d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.001546] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.001800] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.002064] env[65107]: INFO nova.compute.manager [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Shelving [ 966.184241] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103178, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.201508] env[65107]: DEBUG nova.scheduler.client.report [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 966.302644] env[65107]: WARNING openstack [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 966.303255] env[65107]: WARNING openstack [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 966.391034] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103180, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.426788] env[65107]: INFO nova.compute.manager [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Took 30.83 seconds to build instance. [ 966.588426] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 966.623919] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 966.624208] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 966.624339] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 966.624520] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 966.624660] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 966.624805] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 966.625071] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.625233] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 966.625418] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 966.625582] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 966.625804] env[65107]: DEBUG nova.virt.hardware [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 966.626779] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43383f2-abb6-4597-85d8-440488da53ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.636785] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e2d06f-2049-4521-89df-22cd298ce381 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.683242] env[65107]: DEBUG oslo_vmware.api [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103178, 'name': PowerOnVM_Task, 'duration_secs': 1.136594} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.683549] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.684208] env[65107]: INFO nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Took 17.75 seconds to spawn the instance on the hypervisor. [ 966.684208] env[65107]: DEBUG nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 966.685038] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ae9931-1201-4ce7-a70a-159bdbc13f82 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.709068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.149s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.713253] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.702s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.713253] env[65107]: DEBUG nova.objects.instance [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 966.744474] env[65107]: WARNING neutronclient.v2_0.client [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 966.745276] env[65107]: WARNING openstack [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 966.745668] env[65107]: WARNING openstack [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 966.759135] env[65107]: INFO nova.scheduler.client.report [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Deleted allocations for instance cddb14db-34c3-4783-b21e-e0b4623a3a9e [ 966.897916] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103180, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.917634] env[65107]: DEBUG nova.network.neutron [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updated VIF entry in instance network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 966.918073] env[65107]: DEBUG nova.network.neutron [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 966.929031] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6bd3c151-159c-4469-b281-3d2547cb54ef tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "c5fb4b1c-5540-4645-b996-f397729beb23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.345s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.989899] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.990148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.013223] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.013528] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f51c1300-f5e6-44c6-b690-7e1dcf53b5aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.023271] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 967.023271] env[65107]: value = "task-5103181" [ 967.023271] env[65107]: _type = "Task" [ 967.023271] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.032393] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103181, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.207510] env[65107]: INFO nova.compute.manager [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Took 47.66 seconds to build instance. [ 967.269943] env[65107]: DEBUG nova.network.neutron [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Successfully updated port: 69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 967.272090] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eb21569a-d821-4fd7-8c3b-e63a6a4a123a tempest-ServersTestFqdnHostnames-624065465 tempest-ServersTestFqdnHostnames-624065465-project-member] Lock "cddb14db-34c3-4783-b21e-e0b4623a3a9e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.083s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.398912] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103180, 'name': Rename_Task, 'duration_secs': 1.137373} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.400754] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.400754] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85b41df9-65bd-49ee-95c5-daf82ed1b96d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.408207] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 967.408207] env[65107]: value = "task-5103182" [ 967.408207] env[65107]: _type = "Task" [ 967.408207] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.420018] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.425796] env[65107]: DEBUG oslo_concurrency.lockutils [req-9db0a887-c53e-48b1-9b6c-77d70050c79b req-888d1bad-379d-4307-b9a1-4e89a46ff93b service nova] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 967.434954] env[65107]: DEBUG nova.compute.manager [req-dee08936-1f1c-41c8-a210-52c1d1e07e67 req-46bf961a-4ee3-41bc-b73a-7396cfc37d39 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Received event network-vif-plugged-69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 967.435208] env[65107]: DEBUG oslo_concurrency.lockutils [req-dee08936-1f1c-41c8-a210-52c1d1e07e67 req-46bf961a-4ee3-41bc-b73a-7396cfc37d39 service nova] Acquiring lock "f4b1352b-1c55-4987-a298-69431c4e565d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.435440] env[65107]: DEBUG oslo_concurrency.lockutils [req-dee08936-1f1c-41c8-a210-52c1d1e07e67 req-46bf961a-4ee3-41bc-b73a-7396cfc37d39 service nova] Lock "f4b1352b-1c55-4987-a298-69431c4e565d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.435600] env[65107]: DEBUG oslo_concurrency.lockutils [req-dee08936-1f1c-41c8-a210-52c1d1e07e67 req-46bf961a-4ee3-41bc-b73a-7396cfc37d39 service nova] Lock "f4b1352b-1c55-4987-a298-69431c4e565d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.435733] env[65107]: DEBUG nova.compute.manager [req-dee08936-1f1c-41c8-a210-52c1d1e07e67 req-46bf961a-4ee3-41bc-b73a-7396cfc37d39 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] No waiting events found dispatching network-vif-plugged-69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 967.436096] env[65107]: WARNING nova.compute.manager [req-dee08936-1f1c-41c8-a210-52c1d1e07e67 req-46bf961a-4ee3-41bc-b73a-7396cfc37d39 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Received unexpected event network-vif-plugged-69598f61-761f-4866-8b1d-561df2ebbe70 for instance with vm_state building and task_state spawning. [ 967.493222] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 967.536610] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103181, 'name': PowerOffVM_Task, 'duration_secs': 0.258777} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.537756] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 967.538050] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca37498b-91e0-4a6f-9d9e-c7349e424ed9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.566054] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee66556d-4d6c-4f4d-9d7a-ac06e85ff581 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.710164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-20dabf20-901f-4fa2-9b79-cccae7df58f8 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.180s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.727137] env[65107]: DEBUG oslo_concurrency.lockutils [None req-80d35e79-a163-4dc3-bc93-d4fde8fadcd2 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.728309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.201s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.730866] env[65107]: INFO nova.compute.claims [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.771156] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "refresh_cache-f4b1352b-1c55-4987-a298-69431c4e565d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.771394] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquired lock "refresh_cache-f4b1352b-1c55-4987-a298-69431c4e565d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 967.771582] env[65107]: DEBUG nova.network.neutron [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 967.919328] env[65107]: DEBUG oslo_vmware.api [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103182, 'name': PowerOnVM_Task, 'duration_secs': 0.471987} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.919621] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.919820] env[65107]: INFO nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Took 5.88 seconds to spawn the instance on the hypervisor. [ 967.919999] env[65107]: DEBUG nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 967.920808] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f1c4d5-5a7b-4b00-b193-19b93c3569eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.018877] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.073744] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.074200] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.074528] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.074757] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.075018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 968.078158] env[65107]: INFO nova.compute.manager [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Terminating instance [ 968.081679] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 968.081974] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d10ed8de-763e-4aa7-b74a-3f14c6cbf071 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.092107] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 968.092107] env[65107]: value = "task-5103183" [ 968.092107] env[65107]: _type = "Task" [ 968.092107] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.104173] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103183, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.275450] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 968.275717] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 968.350771] env[65107]: DEBUG nova.network.neutron [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 968.392873] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 968.393065] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 968.448039] env[65107]: INFO nova.compute.manager [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Took 32.47 seconds to build instance. [ 968.481417] env[65107]: WARNING neutronclient.v2_0.client [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 968.486668] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 968.486668] env[65107]: WARNING openstack [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 968.578775] env[65107]: DEBUG nova.network.neutron [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Updating instance_info_cache with network_info: [{"id": "69598f61-761f-4866-8b1d-561df2ebbe70", "address": "fa:16:3e:c4:b9:25", "network": {"id": "200f24ea-f070-462f-95ee-0241d4ed2999", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1440594689-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "746d58fc48e64fc58ac57ad38b8051f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69598f61-76", "ovs_interfaceid": "69598f61-761f-4866-8b1d-561df2ebbe70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 968.582345] env[65107]: DEBUG nova.compute.manager [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 968.582598] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.583487] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b77dcd9-9b13-4e41-b832-d62adb80c726 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.595322] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.600621] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc9f31ef-1b6d-44c6-ab0e-d8210ef138d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.612637] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103183, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.614222] env[65107]: DEBUG oslo_vmware.api [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 968.614222] env[65107]: value = "task-5103184" [ 968.614222] env[65107]: _type = "Task" [ 968.614222] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.626858] env[65107]: DEBUG oslo_vmware.api [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.950960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-dc1f3a90-fb1c-4e5a-8de7-cf7ce769707f tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "58345821-536a-46ad-af55-92de7552e924" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.985s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.081409] env[65107]: INFO nova.compute.manager [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Rebuilding instance [ 969.086537] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Releasing lock "refresh_cache-f4b1352b-1c55-4987-a298-69431c4e565d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 969.086951] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Instance network_info: |[{"id": "69598f61-761f-4866-8b1d-561df2ebbe70", "address": "fa:16:3e:c4:b9:25", "network": {"id": "200f24ea-f070-462f-95ee-0241d4ed2999", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1440594689-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "746d58fc48e64fc58ac57ad38b8051f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69598f61-76", "ovs_interfaceid": "69598f61-761f-4866-8b1d-561df2ebbe70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 969.090333] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:b9:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2c424c9-6446-4b2a-af8c-4d9c29117c39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69598f61-761f-4866-8b1d-561df2ebbe70', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.098276] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Creating folder: Project (746d58fc48e64fc58ac57ad38b8051f3). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.099208] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be58a53a-1a78-4a19-9008-9393624e9658 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.113334] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103183, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.124138] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Created folder: Project (746d58fc48e64fc58ac57ad38b8051f3) in parent group-v992574. [ 969.125072] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Creating folder: Instances. Parent ref: group-v992808. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.129025] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-525416a0-7295-4af9-b7ec-53885ddd6767 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.140986] env[65107]: DEBUG oslo_vmware.api [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103184, 'name': PowerOffVM_Task, 'duration_secs': 0.398391} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.141326] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.141501] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.141760] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0457f073-3527-435e-b178-ac724e90fc0b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.146875] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Created folder: Instances in parent group-v992808. [ 969.147200] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 969.147355] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.147576] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81f05950-de63-493f-8023-aff6b45b044b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.174462] env[65107]: DEBUG nova.compute.manager [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 969.174835] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2de13b4-3b10-4101-ac9a-6514157e03fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.183726] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.183726] env[65107]: value = "task-5103188" [ 969.183726] env[65107]: _type = "Task" [ 969.183726] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.200043] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103188, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.232468] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.233727] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.233727] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleting the datastore file [datastore1] 52bd7ecb-4f53-443e-8916-3ff50dfab2b6 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.235113] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5788f301-9c07-498c-af32-69f6ce99ed60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.238416] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a177ba-09e1-41fc-9cc9-cb0e3d1d4d04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.251793] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe0f6d9-b8a1-42e0-95af-1c131f81259c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.256137] env[65107]: DEBUG oslo_vmware.api [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 969.256137] env[65107]: value = "task-5103189" [ 969.256137] env[65107]: _type = "Task" [ 969.256137] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.291240] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de71159-c240-48f0-9325-23cc009cc239 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.297674] env[65107]: DEBUG oslo_vmware.api [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103189, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.303715] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5db7ae-df97-4145-9c2d-e55fde6e3a3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.319172] env[65107]: DEBUG nova.compute.provider_tree [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.534176] env[65107]: DEBUG nova.compute.manager [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Received event network-changed-69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 969.534445] env[65107]: DEBUG nova.compute.manager [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Refreshing instance network info cache due to event network-changed-69598f61-761f-4866-8b1d-561df2ebbe70. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 969.536045] env[65107]: DEBUG oslo_concurrency.lockutils [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Acquiring lock "refresh_cache-f4b1352b-1c55-4987-a298-69431c4e565d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.536045] env[65107]: DEBUG oslo_concurrency.lockutils [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Acquired lock "refresh_cache-f4b1352b-1c55-4987-a298-69431c4e565d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 969.536045] env[65107]: DEBUG nova.network.neutron [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Refreshing network info cache for port 69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 969.609638] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103183, 'name': CreateSnapshot_Task, 'duration_secs': 1.080786} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.610092] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 969.610690] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d830a47-bd8c-4119-9c34-037d2829d12a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.695646] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103188, 'name': CreateVM_Task, 'duration_secs': 0.476626} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.695936] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 969.696618] env[65107]: WARNING neutronclient.v2_0.client [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.697243] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.697243] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 969.697522] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 969.697826] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a813bc00-f57b-4def-8665-9636cd912784 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.705516] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 969.705516] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e485cf-4673-d972-9084-5a0106f2add8" [ 969.705516] env[65107]: _type = "Task" [ 969.705516] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.715929] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e485cf-4673-d972-9084-5a0106f2add8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.766085] env[65107]: DEBUG oslo_vmware.api [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103189, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266403} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.766389] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.766579] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.766835] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.766994] env[65107]: INFO nova.compute.manager [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 969.768504] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 969.768504] env[65107]: DEBUG nova.compute.manager [-] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 969.768504] env[65107]: DEBUG nova.network.neutron [-] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 969.768504] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.768504] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.768776] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.808286] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.822711] env[65107]: DEBUG nova.scheduler.client.report [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 970.039181] env[65107]: WARNING neutronclient.v2_0.client [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.040245] env[65107]: WARNING openstack [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.040749] env[65107]: WARNING openstack [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.131464] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 970.132319] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-739c7ff2-cae7-4f40-aa3e-5df2535a8d3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.142922] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 970.142922] env[65107]: value = "task-5103190" [ 970.142922] env[65107]: _type = "Task" [ 970.142922] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.156322] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103190, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.190649] env[65107]: WARNING openstack [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.190649] env[65107]: WARNING openstack [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.202988] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.203694] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86545e5a-566c-49cd-85d0-3ea6ac821766 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.214543] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 970.214543] env[65107]: value = "task-5103191" [ 970.214543] env[65107]: _type = "Task" [ 970.214543] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.220051] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e485cf-4673-d972-9084-5a0106f2add8, 'name': SearchDatastore_Task, 'duration_secs': 0.013781} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.223606] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.223911] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.224208] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.224376] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.224591] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.224979] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b609bc2-8aad-400c-9dd9-69c54a919da7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.237641] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103191, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.242763] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.242906] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.243771] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d237604c-7e12-4b3e-9472-ebfa2b2bc5b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.251476] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 970.251476] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae3403-3707-6ce1-7f09-d39584bce4d0" [ 970.251476] env[65107]: _type = "Task" [ 970.251476] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.261416] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae3403-3707-6ce1-7f09-d39584bce4d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.271138] env[65107]: WARNING neutronclient.v2_0.client [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.271824] env[65107]: WARNING openstack [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.272219] env[65107]: WARNING openstack [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.328422] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.328953] env[65107]: DEBUG nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 970.335862] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.375s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.337735] env[65107]: INFO nova.compute.claims [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.402862] env[65107]: DEBUG nova.network.neutron [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Updated VIF entry in instance network info cache for port 69598f61-761f-4866-8b1d-561df2ebbe70. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 970.403256] env[65107]: DEBUG nova.network.neutron [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Updating instance_info_cache with network_info: [{"id": "69598f61-761f-4866-8b1d-561df2ebbe70", "address": "fa:16:3e:c4:b9:25", "network": {"id": "200f24ea-f070-462f-95ee-0241d4ed2999", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1440594689-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "746d58fc48e64fc58ac57ad38b8051f3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69598f61-76", "ovs_interfaceid": "69598f61-761f-4866-8b1d-561df2ebbe70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.554181] env[65107]: DEBUG nova.network.neutron [-] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.653910] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103190, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.730055] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103191, 'name': PowerOffVM_Task, 'duration_secs': 0.146504} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.730111] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.731375] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.731626] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2291d61b-1bf6-44c5-b04b-667365249a02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.740853] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.740853] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7169f25d-ca39-45a8-b515-dbc1c7185926 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.762264] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ae3403-3707-6ce1-7f09-d39584bce4d0, 'name': SearchDatastore_Task, 'duration_secs': 0.015621} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.763285] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e226d234-71f8-4080-a057-37c1b36f2116 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.767577] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.767789] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.767967] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleting the datastore file [datastore2] 58345821-536a-46ad-af55-92de7552e924 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.768282] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b46a0b5-3719-4cc8-9398-0311105486ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.771716] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 970.771716] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de5499-2fd9-24ee-8903-121fa90797e5" [ 970.771716] env[65107]: _type = "Task" [ 970.771716] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.780508] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 970.780508] env[65107]: value = "task-5103193" [ 970.780508] env[65107]: _type = "Task" [ 970.780508] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.787450] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de5499-2fd9-24ee-8903-121fa90797e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.793275] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103193, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.843753] env[65107]: DEBUG nova.compute.utils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 970.848068] env[65107]: DEBUG nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 970.905861] env[65107]: DEBUG oslo_concurrency.lockutils [req-75904650-d52b-48e2-9a92-3b64eff04bc2 req-121f8fad-2f4c-4b51-aa0a-273e56983135 service nova] Releasing lock "refresh_cache-f4b1352b-1c55-4987-a298-69431c4e565d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.056226] env[65107]: INFO nova.compute.manager [-] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Took 1.29 seconds to deallocate network for instance. [ 971.158263] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103190, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.286082] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de5499-2fd9-24ee-8903-121fa90797e5, 'name': SearchDatastore_Task, 'duration_secs': 0.020832} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.289750] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.290177] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] f4b1352b-1c55-4987-a298-69431c4e565d/f4b1352b-1c55-4987-a298-69431c4e565d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 971.290522] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dddc5471-3cda-447b-a73e-4f2d72d27766 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.298822] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103193, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.271173} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.300793] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.301074] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.301284] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.303930] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 971.303930] env[65107]: value = "task-5103194" [ 971.303930] env[65107]: _type = "Task" [ 971.303930] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.312542] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103194, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.349181] env[65107]: DEBUG nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 971.565264] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.566839] env[65107]: DEBUG nova.compute.manager [req-2be79c71-f07e-47e9-94fe-aeb4644b419a req-fe5fd715-75e9-4f87-87f8-73920c8ad25f service nova] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Received event network-vif-deleted-853b4d7e-c8ee-4570-a59c-310c21f06168 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 971.654158] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103190, 'name': CloneVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.745242] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238ceb2b-8ee9-4235-af83-c98da863ce3a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.754470] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4f0053-d9cd-46e7-8164-2639a7531f1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.788038] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a6c087-4da8-4702-9485-2b6dfbbd1d5a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.799426] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a031e3-d287-48a2-b8c1-d0b04f6074ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.832120] env[65107]: DEBUG nova.compute.provider_tree [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.840338] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103194, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.155567] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103190, 'name': CloneVM_Task, 'duration_secs': 1.596996} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.155918] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Created linked-clone VM from snapshot [ 972.156666] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f852485-72ba-404a-b9f4-72b713684845 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.165560] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Uploading image c38e9e3f-de87-4115-b1da-c44beabd8098 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 972.191138] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 972.191138] env[65107]: value = "vm-992812" [ 972.191138] env[65107]: _type = "VirtualMachine" [ 972.191138] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 972.191440] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e7b0a584-77d8-4351-aeb4-0b68afca1991 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.199478] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease: (returnval){ [ 972.199478] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5265faa2-ceb3-ec50-0f58-bde96b1cfc45" [ 972.199478] env[65107]: _type = "HttpNfcLease" [ 972.199478] env[65107]: } obtained for exporting VM: (result){ [ 972.199478] env[65107]: value = "vm-992812" [ 972.199478] env[65107]: _type = "VirtualMachine" [ 972.199478] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 972.199478] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the lease: (returnval){ [ 972.199478] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5265faa2-ceb3-ec50-0f58-bde96b1cfc45" [ 972.199478] env[65107]: _type = "HttpNfcLease" [ 972.199478] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 972.206510] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 972.206510] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5265faa2-ceb3-ec50-0f58-bde96b1cfc45" [ 972.206510] env[65107]: _type = "HttpNfcLease" [ 972.206510] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 972.316950] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103194, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.78865} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.317325] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] f4b1352b-1c55-4987-a298-69431c4e565d/f4b1352b-1c55-4987-a298-69431c4e565d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 972.317589] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.317819] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dacf35cc-495f-4bf7-a43a-0ad783d4d0a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.325829] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 972.325829] env[65107]: value = "task-5103196" [ 972.325829] env[65107]: _type = "Task" [ 972.325829] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.339201] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103196, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.341955] env[65107]: DEBUG nova.scheduler.client.report [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 972.362801] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 972.363137] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 972.363302] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 972.363488] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 972.363634] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 972.363780] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 972.364078] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.364282] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 972.364481] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 972.364641] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 972.364814] env[65107]: DEBUG nova.virt.hardware [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 972.366392] env[65107]: DEBUG nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 972.369057] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9171ccee-a56c-4a27-84b1-5c3d54f33ee5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.379650] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36febf6-a872-4a97-b5fa-9c597d4ee8b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.394449] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.400095] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 972.402416] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 972.402641] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 972.402794] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 972.402980] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 972.403135] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 972.403285] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 972.403489] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.403643] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 972.403801] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 972.403961] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 972.404188] env[65107]: DEBUG nova.virt.hardware [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 972.404781] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 972.405508] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42e0726-9eb4-4655-9a8f-7b969dd8049e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.408100] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09a6fbed-dec2-4fa5-b600-20df4d98a33f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.427827] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af587b94-16fa-4662-b1bc-df74517de072 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.431692] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.431692] env[65107]: value = "task-5103197" [ 972.431692] env[65107]: _type = "Task" [ 972.431692] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.444902] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.450615] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Creating folder: Project (369b36eb93a94c33b598e7ebfc2f7099). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 972.451424] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb5c0e96-8d4b-4352-9059-3a9375b80631 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.456225] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103197, 'name': CreateVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.465698] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Created folder: Project (369b36eb93a94c33b598e7ebfc2f7099) in parent group-v992574. [ 972.465869] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Creating folder: Instances. Parent ref: group-v992814. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 972.466174] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6e1965d-5a46-4aaa-9a30-05bbeb6e6c16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.476886] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Created folder: Instances in parent group-v992814. [ 972.477117] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 972.477346] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 972.477562] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff508e10-0c20-4ae4-a13e-91b8bbf8748f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.496561] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.496561] env[65107]: value = "task-5103200" [ 972.496561] env[65107]: _type = "Task" [ 972.496561] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.505403] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103200, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.708562] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 972.708562] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5265faa2-ceb3-ec50-0f58-bde96b1cfc45" [ 972.708562] env[65107]: _type = "HttpNfcLease" [ 972.708562] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 972.708881] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 972.708881] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5265faa2-ceb3-ec50-0f58-bde96b1cfc45" [ 972.708881] env[65107]: _type = "HttpNfcLease" [ 972.708881] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 972.709680] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf0efe6-2265-42f5-9c5b-d854fa3b6f59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.719449] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a58be0-c980-09e9-a375-6438f3ef29c6/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 972.719827] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a58be0-c980-09e9-a375-6438f3ef29c6/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 972.811188] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-61b9a847-2ab1-43c5-b92d-8243b0472e19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.840089] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103196, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.21672} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.846723] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.847790] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.848316] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 972.851803] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a89b87-724a-4cda-8081-a54285168e2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.856349] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.408s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.856349] env[65107]: DEBUG nova.objects.instance [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lazy-loading 'resources' on Instance uuid 40474981-eeef-492e-8a8b-aaea7f554c02 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.880440] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] f4b1352b-1c55-4987-a298-69431c4e565d/f4b1352b-1c55-4987-a298-69431c4e565d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.881182] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-504e19ff-7eb2-4bb2-a224-7acea638cc8d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.904533] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 972.904533] env[65107]: value = "task-5103201" [ 972.904533] env[65107]: _type = "Task" [ 972.904533] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.914492] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103201, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.942661] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103197, 'name': CreateVM_Task, 'duration_secs': 0.351962} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.942790] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 972.943361] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.943569] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.943929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 972.944245] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63c969c9-7bf5-4631-a72d-ea44574c088d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.950226] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 972.950226] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c1a24-393b-bd51-f176-5e5843e6263f" [ 972.950226] env[65107]: _type = "Task" [ 972.950226] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.959670] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c1a24-393b-bd51-f176-5e5843e6263f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.007787] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103200, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.357428] env[65107]: DEBUG nova.compute.utils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 973.359204] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 973.360047] env[65107]: DEBUG nova.network.neutron [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 973.360900] env[65107]: WARNING neutronclient.v2_0.client [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.360900] env[65107]: WARNING neutronclient.v2_0.client [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.361950] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.362506] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.416954] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103201, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.431718] env[65107]: DEBUG nova.policy [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '365eed31877241e58fc860ed5620d5f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c805fe68d0fd4147b9b4b7fdc9d6a6d0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 973.467550] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c1a24-393b-bd51-f176-5e5843e6263f, 'name': SearchDatastore_Task, 'duration_secs': 0.029093} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.467876] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 973.468141] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 973.468481] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.468646] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 973.468831] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 973.469137] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-251b0d45-e01c-4dc0-aa79-4a5c3b7cd398 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.480036] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 973.480249] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 973.481018] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be12b96d-1709-4c5c-9e37-49b6ba1e6a55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.493913] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 973.493913] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5f635-a6ed-29d4-58ad-376e91c680f6" [ 973.493913] env[65107]: _type = "Task" [ 973.493913] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.505815] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5f635-a6ed-29d4-58ad-376e91c680f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.515230] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103200, 'name': CreateVM_Task, 'duration_secs': 0.597174} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.515511] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 973.516066] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.516290] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 973.516805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 973.517107] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2cde156-b702-466e-b278-2c8f9030f80f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.522490] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 973.522490] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcbd26-2d7c-db67-6590-dd1cd0e6d60e" [ 973.522490] env[65107]: _type = "Task" [ 973.522490] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.535040] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcbd26-2d7c-db67-6590-dd1cd0e6d60e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.771352] env[65107]: DEBUG nova.network.neutron [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Successfully created port: a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 973.793885] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463ee227-9d8a-4e67-8303-4c9a05d0c465 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.803379] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428c2575-9950-4c9c-b72c-23ab25ebb58a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.837322] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03102436-81cc-49c4-95e4-750263e457cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.846080] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a705f7e-8610-4734-a296-d9d0a5dbb4b3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.862433] env[65107]: DEBUG nova.compute.provider_tree [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.874600] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 973.916697] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103201, 'name': ReconfigVM_Task, 'duration_secs': 0.644801} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.917213] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Reconfigured VM instance instance-00000053 to attach disk [datastore1] f4b1352b-1c55-4987-a298-69431c4e565d/f4b1352b-1c55-4987-a298-69431c4e565d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.917928] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3a40f9d-c8bf-43fa-bdcf-695b2d4b51ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.925411] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 973.925411] env[65107]: value = "task-5103202" [ 973.925411] env[65107]: _type = "Task" [ 973.925411] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.939063] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103202, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.008832] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5f635-a6ed-29d4-58ad-376e91c680f6, 'name': SearchDatastore_Task, 'duration_secs': 0.014174} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.009818] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0da11f8-9af0-420d-b47a-6f5207a7b61d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.018709] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 974.018709] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c12bc3-d8cb-8f70-974a-0541232cfddd" [ 974.018709] env[65107]: _type = "Task" [ 974.018709] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.027957] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c12bc3-d8cb-8f70-974a-0541232cfddd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.036526] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcbd26-2d7c-db67-6590-dd1cd0e6d60e, 'name': SearchDatastore_Task, 'duration_secs': 0.013617} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.036991] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.037260] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.038864] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.367060] env[65107]: DEBUG nova.scheduler.client.report [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 974.436978] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103202, 'name': Rename_Task, 'duration_secs': 0.154941} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.437381] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.437727] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6041976a-da53-4cb9-a1a4-3891313c7b9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.445570] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 974.445570] env[65107]: value = "task-5103203" [ 974.445570] env[65107]: _type = "Task" [ 974.445570] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.455932] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103203, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.527032] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c12bc3-d8cb-8f70-974a-0541232cfddd, 'name': SearchDatastore_Task, 'duration_secs': 0.010619} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.527354] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.527708] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.528069] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 974.528388] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.528656] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3d600c4-ba06-45f5-ab7e-bfa382fd88ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.530827] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07ad0293-1dc7-403c-9a09-c8be8db019a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.539701] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 974.539701] env[65107]: value = "task-5103204" [ 974.539701] env[65107]: _type = "Task" [ 974.539701] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.541303] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.541497] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.545726] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f33968c-555a-4d48-a2f9-d3f66f82b9c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.554242] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.555723] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 974.555723] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f556a5-6c68-685a-8cf8-fd8d2dbc008e" [ 974.555723] env[65107]: _type = "Task" [ 974.555723] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.564054] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f556a5-6c68-685a-8cf8-fd8d2dbc008e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.872244] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.017s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.875237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.240s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.875534] env[65107]: DEBUG nova.objects.instance [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'resources' on Instance uuid fead9fe9-23f4-4d13-bf44-be0727057ddd {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.885639] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 974.911268] env[65107]: INFO nova.scheduler.client.report [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Deleted allocations for instance 40474981-eeef-492e-8a8b-aaea7f554c02 [ 974.924241] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 974.924241] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 974.924241] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 974.924936] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 974.925216] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 974.925395] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 974.925634] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.925805] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 974.925979] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 974.926198] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 974.926412] env[65107]: DEBUG nova.virt.hardware [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 974.927806] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf5d185-d171-4d68-8ecd-da59befb7197 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.950685] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f674d6d-8a0d-4b69-ae0c-2a1534f28661 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.971533] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103203, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.057937] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103204, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.072800] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f556a5-6c68-685a-8cf8-fd8d2dbc008e, 'name': SearchDatastore_Task, 'duration_secs': 0.018676} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.074262] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a647915-68e6-44f2-aa54-cd323f4e7642 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.082205] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 975.082205] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b78db8-510f-326c-80b9-11749768469d" [ 975.082205] env[65107]: _type = "Task" [ 975.082205] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.094736] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b78db8-510f-326c-80b9-11749768469d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.365129] env[65107]: DEBUG nova.compute.manager [req-6dcf1acc-179d-4a37-abf3-5bbebae6231c req-2d0be204-c2db-44e0-be60-58aae17385f8 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Received event network-vif-plugged-a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 975.365499] env[65107]: DEBUG oslo_concurrency.lockutils [req-6dcf1acc-179d-4a37-abf3-5bbebae6231c req-2d0be204-c2db-44e0-be60-58aae17385f8 service nova] Acquiring lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 975.365625] env[65107]: DEBUG oslo_concurrency.lockutils [req-6dcf1acc-179d-4a37-abf3-5bbebae6231c req-2d0be204-c2db-44e0-be60-58aae17385f8 service nova] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.365794] env[65107]: DEBUG oslo_concurrency.lockutils [req-6dcf1acc-179d-4a37-abf3-5bbebae6231c req-2d0be204-c2db-44e0-be60-58aae17385f8 service nova] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.365967] env[65107]: DEBUG nova.compute.manager [req-6dcf1acc-179d-4a37-abf3-5bbebae6231c req-2d0be204-c2db-44e0-be60-58aae17385f8 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] No waiting events found dispatching network-vif-plugged-a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 975.366155] env[65107]: WARNING nova.compute.manager [req-6dcf1acc-179d-4a37-abf3-5bbebae6231c req-2d0be204-c2db-44e0-be60-58aae17385f8 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Received unexpected event network-vif-plugged-a34ba705-83e3-4da1-ad09-b23e6fc8d54e for instance with vm_state building and task_state spawning. [ 975.415729] env[65107]: DEBUG nova.network.neutron [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Successfully updated port: a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 975.425728] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3a4d5600-00b1-4088-83b6-e84cea78f0c0 tempest-ListServerFiltersTestJSON-1901090165 tempest-ListServerFiltersTestJSON-1901090165-project-member] Lock "40474981-eeef-492e-8a8b-aaea7f554c02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.943s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.471523] env[65107]: DEBUG oslo_vmware.api [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103203, 'name': PowerOnVM_Task, 'duration_secs': 0.86829} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.471971] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.471971] env[65107]: INFO nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Took 8.88 seconds to spawn the instance on the hypervisor. [ 975.474804] env[65107]: DEBUG nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 975.475647] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61606b56-1626-4ed1-a5f9-4c32062453ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.552402] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103204, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559081} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.555125] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 975.555947] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.556472] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-386ef740-94e7-4957-afa6-89886f72a319 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.564278] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 975.564278] env[65107]: value = "task-5103205" [ 975.564278] env[65107]: _type = "Task" [ 975.564278] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.580851] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103205, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.598398] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b78db8-510f-326c-80b9-11749768469d, 'name': SearchDatastore_Task, 'duration_secs': 0.01224} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.598398] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 975.598663] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.598958] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14168ab5-b06a-458b-9b66-0061fea8cac4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.607073] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 975.607073] env[65107]: value = "task-5103206" [ 975.607073] env[65107]: _type = "Task" [ 975.607073] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.616518] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.770587] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a0c8cf-839f-4cc0-8e0a-87b25db3887b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.778648] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159d1280-17ed-4929-97fb-e058d4dea34e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.812667] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03602c8-d95a-4278-a7ab-abd3049abe05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.821048] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b625982-5229-43f9-a8e4-f64e40cc18c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.837092] env[65107]: DEBUG nova.compute.provider_tree [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.919517] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.919673] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.919795] env[65107]: DEBUG nova.network.neutron [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 976.000474] env[65107]: INFO nova.compute.manager [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Took 35.55 seconds to build instance. [ 976.079741] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103205, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070937} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.080232] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.081335] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7599317-09b7-410b-a576-703dc6026ecd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.109774] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.111103] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b25d5e65-c727-49af-9593-363374c03cef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.136070] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103206, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.137989] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 976.137989] env[65107]: value = "task-5103207" [ 976.137989] env[65107]: _type = "Task" [ 976.137989] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.148933] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103207, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.340941] env[65107]: DEBUG nova.scheduler.client.report [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 976.422912] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.423243] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.464441] env[65107]: DEBUG nova.network.neutron [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 976.487285] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.487765] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.503284] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5ebf757e-aef6-49b3-b03a-3937d99e8344 tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "f4b1352b-1c55-4987-a298-69431c4e565d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.067s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.557195] env[65107]: WARNING neutronclient.v2_0.client [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 976.557902] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 976.558280] env[65107]: WARNING openstack [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 976.622223] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103206, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637261} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.622585] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.622790] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.623083] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9669072f-529c-454e-9950-d0b4cd4f76ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.631934] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 976.631934] env[65107]: value = "task-5103208" [ 976.631934] env[65107]: _type = "Task" [ 976.631934] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.645971] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.648644] env[65107]: DEBUG nova.network.neutron [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Updating instance_info_cache with network_info: [{"id": "a34ba705-83e3-4da1-ad09-b23e6fc8d54e", "address": "fa:16:3e:0e:8b:83", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34ba705-83", "ovs_interfaceid": "a34ba705-83e3-4da1-ad09-b23e6fc8d54e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 976.659172] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103207, 'name': ReconfigVM_Task, 'duration_secs': 0.320487} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.659602] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 58345821-536a-46ad-af55-92de7552e924/58345821-536a-46ad-af55-92de7552e924.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.661063] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0b96166-37f1-4bee-9691-a3c55097c845 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.670895] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 976.670895] env[65107]: value = "task-5103209" [ 976.670895] env[65107]: _type = "Task" [ 976.670895] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.686030] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103209, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.846792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.972s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.849946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.431s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.850417] env[65107]: DEBUG nova.objects.instance [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'resources' on Instance uuid 3f34d6a2-0370-4239-9556-b4333e454d58 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.871058] env[65107]: INFO nova.scheduler.client.report [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocations for instance fead9fe9-23f4-4d13-bf44-be0727057ddd [ 977.143979] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073278} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.144256] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.145213] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9b7f71-8b8e-42bc-b421-f89a23513fb5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.159383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.159762] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance network_info: |[{"id": "a34ba705-83e3-4da1-ad09-b23e6fc8d54e", "address": "fa:16:3e:0e:8b:83", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34ba705-83", "ovs_interfaceid": "a34ba705-83e3-4da1-ad09-b23e6fc8d54e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 977.168884] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.169560] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:8b:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a34ba705-83e3-4da1-ad09-b23e6fc8d54e', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.177104] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 977.177356] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e55cec4-0481-4932-ac5a-9df1a1fd8ee7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.192246] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.196481] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d0ce58e-ce45-43e6-96ed-9d042c6dcf67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.218331] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103209, 'name': Rename_Task, 'duration_secs': 0.163289} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.221034] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.221410] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 977.221410] env[65107]: value = "task-5103210" [ 977.221410] env[65107]: _type = "Task" [ 977.221410] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.221578] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.221578] env[65107]: value = "task-5103211" [ 977.221578] env[65107]: _type = "Task" [ 977.221578] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.221759] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-821a8d4c-bd8a-4fe4-b735-b119bad8676b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.237970] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103211, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.242464] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 977.242464] env[65107]: value = "task-5103212" [ 977.242464] env[65107]: _type = "Task" [ 977.242464] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.242690] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103210, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.252480] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103212, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.324408] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "f4b1352b-1c55-4987-a298-69431c4e565d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.324747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "f4b1352b-1c55-4987-a298-69431c4e565d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.325059] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "f4b1352b-1c55-4987-a298-69431c4e565d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.325257] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "f4b1352b-1c55-4987-a298-69431c4e565d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.325433] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "f4b1352b-1c55-4987-a298-69431c4e565d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.327958] env[65107]: INFO nova.compute.manager [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Terminating instance [ 977.382091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e0b4f7ac-b806-495e-b7e8-c7964fdee0da tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "fead9fe9-23f4-4d13-bf44-be0727057ddd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.005s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.412166] env[65107]: DEBUG nova.compute.manager [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Received event network-changed-a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 977.412299] env[65107]: DEBUG nova.compute.manager [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Refreshing instance network info cache due to event network-changed-a34ba705-83e3-4da1-ad09-b23e6fc8d54e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 977.412516] env[65107]: DEBUG oslo_concurrency.lockutils [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Acquiring lock "refresh_cache-6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.412659] env[65107]: DEBUG oslo_concurrency.lockutils [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Acquired lock "refresh_cache-6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 977.412820] env[65107]: DEBUG nova.network.neutron [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Refreshing network info cache for port a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 977.710985] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63634d6e-891b-419a-976a-0131eba7cbfd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.719590] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f963b9-ac80-4598-9007-5f78b325be2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.760940] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87213f3b-6cb6-4b28-bdb0-006e3165d150 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.770388] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103211, 'name': CreateVM_Task, 'duration_secs': 0.448361} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.772336] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103210, 'name': ReconfigVM_Task, 'duration_secs': 0.511307} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.772336] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 977.772336] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.773161] env[65107]: WARNING neutronclient.v2_0.client [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 977.773664] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.773836] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 977.774197] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 977.774464] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd20cafd-49e5-4824-943d-0e8cfd1cccb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.781673] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4204d7f-51ef-4f22-993f-0830e192491c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.784748] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103212, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.786689] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34696d68-8ccc-4991-80cd-b8db4c3bee10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.792679] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 977.792679] env[65107]: value = "task-5103213" [ 977.792679] env[65107]: _type = "Task" [ 977.792679] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.807788] env[65107]: DEBUG nova.compute.provider_tree [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.809811] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 977.809811] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b83e1e-88d8-8ee4-638d-583834a05cfb" [ 977.809811] env[65107]: _type = "Task" [ 977.809811] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.818379] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103213, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.825693] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b83e1e-88d8-8ee4-638d-583834a05cfb, 'name': SearchDatastore_Task, 'duration_secs': 0.01683} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.827431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 977.827431] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.827431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.827431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 977.827431] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.827431] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53b8108d-872e-4e42-9055-dfc5b2f4730d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.831713] env[65107]: DEBUG nova.compute.manager [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 977.831956] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 977.832864] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c58f4e-c5b9-4b09-835a-380c93d40d41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.838514] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.838514] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 977.841509] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b160e0a9-eff7-4eba-a1ed-f6621f46ae15 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.843947] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 977.844244] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-787b5eb7-7de7-431b-bed5-19446e870ffd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.851709] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 977.851709] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a8ee8a-092f-dc28-192e-04f1c40ba73c" [ 977.851709] env[65107]: _type = "Task" [ 977.851709] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.853840] env[65107]: DEBUG oslo_vmware.api [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 977.853840] env[65107]: value = "task-5103214" [ 977.853840] env[65107]: _type = "Task" [ 977.853840] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.869396] env[65107]: DEBUG oslo_vmware.api [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103214, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.872866] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a8ee8a-092f-dc28-192e-04f1c40ba73c, 'name': SearchDatastore_Task, 'duration_secs': 0.017549} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.873719] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10e73c5f-a1aa-46d7-bf91-8253142d5ab3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.880893] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 977.880893] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c9a448-cc65-8172-68ed-2b2b81ad26f0" [ 977.880893] env[65107]: _type = "Task" [ 977.880893] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.891189] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c9a448-cc65-8172-68ed-2b2b81ad26f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.915871] env[65107]: WARNING neutronclient.v2_0.client [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 977.916711] env[65107]: WARNING openstack [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 977.917092] env[65107]: WARNING openstack [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.030968] env[65107]: WARNING openstack [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.031543] env[65107]: WARNING openstack [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.109204] env[65107]: WARNING neutronclient.v2_0.client [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 978.109988] env[65107]: WARNING openstack [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.110312] env[65107]: WARNING openstack [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.204211] env[65107]: DEBUG nova.network.neutron [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Updated VIF entry in instance network info cache for port a34ba705-83e3-4da1-ad09-b23e6fc8d54e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 978.204719] env[65107]: DEBUG nova.network.neutron [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Updating instance_info_cache with network_info: [{"id": "a34ba705-83e3-4da1-ad09-b23e6fc8d54e", "address": "fa:16:3e:0e:8b:83", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa34ba705-83", "ovs_interfaceid": "a34ba705-83e3-4da1-ad09-b23e6fc8d54e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 978.266252] env[65107]: DEBUG oslo_vmware.api [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103212, 'name': PowerOnVM_Task, 'duration_secs': 1.007861} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.266648] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.266874] env[65107]: DEBUG nova.compute.manager [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 978.267799] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae230ef-673b-4f57-bb69-044199458e3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.303650] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103213, 'name': Rename_Task, 'duration_secs': 0.200571} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.303945] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.304283] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ef98f41-f77d-4ec4-a464-e2b66f4d0e27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.312079] env[65107]: DEBUG nova.scheduler.client.report [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 978.317163] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 978.317163] env[65107]: value = "task-5103215" [ 978.317163] env[65107]: _type = "Task" [ 978.317163] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.327325] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.369144] env[65107]: DEBUG oslo_vmware.api [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103214, 'name': PowerOffVM_Task, 'duration_secs': 0.206491} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.369436] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.369669] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.369858] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76fcd499-38bf-496b-b8cf-54eaca8eaa60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.392503] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c9a448-cc65-8172-68ed-2b2b81ad26f0, 'name': SearchDatastore_Task, 'duration_secs': 0.014225} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.392767] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 978.393051] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.393373] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-426d6f9a-f060-4b74-8626-a0704f6ff191 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.400445] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 978.400445] env[65107]: value = "task-5103217" [ 978.400445] env[65107]: _type = "Task" [ 978.400445] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.410681] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.449681] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.449977] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.450207] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Deleting the datastore file [datastore1] f4b1352b-1c55-4987-a298-69431c4e565d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.450529] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa0c856a-ea27-44db-b42f-f74204aca730 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.457501] env[65107]: DEBUG oslo_vmware.api [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for the task: (returnval){ [ 978.457501] env[65107]: value = "task-5103218" [ 978.457501] env[65107]: _type = "Task" [ 978.457501] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.467311] env[65107]: DEBUG oslo_vmware.api [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103218, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.708358] env[65107]: DEBUG oslo_concurrency.lockutils [req-51000283-0cff-4008-888d-30a8a00820b5 req-448ebc0a-772d-48e6-9d14-73092e797fa3 service nova] Releasing lock "refresh_cache-6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 978.790130] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.819618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.823963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.857s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.825327] env[65107]: INFO nova.compute.claims [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.841319] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103215, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.867609] env[65107]: INFO nova.scheduler.client.report [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted allocations for instance 3f34d6a2-0370-4239-9556-b4333e454d58 [ 978.914310] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103217, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.974018] env[65107]: DEBUG oslo_vmware.api [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Task: {'id': task-5103218, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318677} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.974018] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.974018] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.974259] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.975458] env[65107]: INFO nova.compute.manager [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 978.975759] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 978.975972] env[65107]: DEBUG nova.compute.manager [-] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 978.978056] env[65107]: DEBUG nova.network.neutron [-] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 978.978056] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 978.978737] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.978737] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 979.020395] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.338483] env[65107]: DEBUG oslo_vmware.api [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103215, 'name': PowerOnVM_Task, 'duration_secs': 0.775883} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.339821] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.339821] env[65107]: INFO nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Took 6.97 seconds to spawn the instance on the hypervisor. [ 979.339821] env[65107]: DEBUG nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 979.340652] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f44b2a-3724-4752-930e-3e3a00f464e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.380261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ecb83048-6785-4606-9848-b7e79a6d43ec tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3f34d6a2-0370-4239-9556-b4333e454d58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.936s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.390339] env[65107]: DEBUG nova.compute.manager [req-17d7a4c5-b230-4d39-9dd2-c0769bf95f42 req-582ee45f-4657-4d12-a144-862d8c9d7a28 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Received event network-vif-deleted-69598f61-761f-4866-8b1d-561df2ebbe70 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 979.390339] env[65107]: INFO nova.compute.manager [req-17d7a4c5-b230-4d39-9dd2-c0769bf95f42 req-582ee45f-4657-4d12-a144-862d8c9d7a28 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Neutron deleted interface 69598f61-761f-4866-8b1d-561df2ebbe70; detaching it from the instance and deleting it from the info cache [ 979.390339] env[65107]: DEBUG nova.network.neutron [req-17d7a4c5-b230-4d39-9dd2-c0769bf95f42 req-582ee45f-4657-4d12-a144-862d8c9d7a28 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 979.412304] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543165} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.412684] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.412912] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.413233] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17bc08bc-93ae-49f1-affd-2bf921919a75 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.421248] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 979.421248] env[65107]: value = "task-5103219" [ 979.421248] env[65107]: _type = "Task" [ 979.421248] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.433774] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103219, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.627168] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "58345821-536a-46ad-af55-92de7552e924" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.627435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "58345821-536a-46ad-af55-92de7552e924" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.627807] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "58345821-536a-46ad-af55-92de7552e924-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.628060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "58345821-536a-46ad-af55-92de7552e924-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.628273] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "58345821-536a-46ad-af55-92de7552e924-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.630559] env[65107]: INFO nova.compute.manager [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Terminating instance [ 979.794090] env[65107]: DEBUG nova.network.neutron [-] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 979.867896] env[65107]: INFO nova.compute.manager [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Took 33.37 seconds to build instance. [ 979.893040] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbc13134-2966-4507-afab-442c37a1d2c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.903063] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1221d75b-0cf1-4eae-9949-9618868d0367 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.932181] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103219, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198502} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.945947] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.947016] env[65107]: DEBUG nova.compute.manager [req-17d7a4c5-b230-4d39-9dd2-c0769bf95f42 req-582ee45f-4657-4d12-a144-862d8c9d7a28 service nova] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Detach interface failed, port_id=69598f61-761f-4866-8b1d-561df2ebbe70, reason: Instance f4b1352b-1c55-4987-a298-69431c4e565d could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 979.950767] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd6178a-f016-4763-a85d-9d30acc38165 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.974677] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.977617] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06b35c01-2254-4086-820f-c8c4103f7ddb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.998746] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 979.998746] env[65107]: value = "task-5103220" [ 979.998746] env[65107]: _type = "Task" [ 979.998746] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.013723] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103220, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.136612] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "refresh_cache-58345821-536a-46ad-af55-92de7552e924" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.136853] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "refresh_cache-58345821-536a-46ad-af55-92de7552e924" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 980.137060] env[65107]: DEBUG nova.network.neutron [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 980.154351] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.154506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.250051] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f701ffed-515f-4fb9-a464-64813429e8ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.258521] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c770869c-2e4c-41d0-a548-cf9edffaf828 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.290968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0816b7fa-84cd-4b4e-b8c5-7d3c0b252d02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.298575] env[65107]: INFO nova.compute.manager [-] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Took 1.32 seconds to deallocate network for instance. [ 980.301512] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473c36ec-02c0-46c7-858a-066ba619b5dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.321980] env[65107]: DEBUG nova.compute.provider_tree [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.370689] env[65107]: DEBUG oslo_concurrency.lockutils [None req-46f7a3f3-7b61-4210-838d-ecceb3f756f5 tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "13c96e2f-46de-46e0-a505-60dfec9e95ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.882s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.510024] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103220, 'name': ReconfigVM_Task, 'duration_secs': 0.464704} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.510596] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.511595] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88d9cae9-da5e-47eb-bc8e-743f0d58f009 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.522270] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 980.522270] env[65107]: value = "task-5103221" [ 980.522270] env[65107]: _type = "Task" [ 980.522270] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.533648] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103221, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.639850] env[65107]: WARNING neutronclient.v2_0.client [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 980.640590] env[65107]: WARNING openstack [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 980.640986] env[65107]: WARNING openstack [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 980.658429] env[65107]: INFO nova.compute.manager [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Detaching volume aae9fa95-1cf2-433e-94d4-b3c0b1119742 [ 980.665545] env[65107]: DEBUG nova.network.neutron [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 980.705867] env[65107]: INFO nova.virt.block_device [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Attempting to driver detach volume aae9fa95-1cf2-433e-94d4-b3c0b1119742 from mountpoint /dev/sdb [ 980.706139] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 980.706361] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992733', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'name': 'volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '82e8c093-b718-4d38-9682-ba2e710e4b93', 'attached_at': '', 'detached_at': '', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'serial': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 980.708526] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b3793a-5e3e-4be3-85ef-dc41344a09b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.735032] env[65107]: DEBUG nova.network.neutron [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.737054] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3178a996-bff9-4aba-b6dd-3706c965a15e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.746574] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52e0447-bbea-469a-93cc-d49855ca4bda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.769433] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c440efe4-028c-4a88-8a3c-a83ed889d35b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.785871] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The volume has not been displaced from its original location: [datastore1] volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742/volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 980.791240] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Reconfiguring VM instance instance-0000002b to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 980.791611] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f94bc171-125f-45e3-952c-86629d711215 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.810929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.812790] env[65107]: DEBUG oslo_vmware.api [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 980.812790] env[65107]: value = "task-5103222" [ 980.812790] env[65107]: _type = "Task" [ 980.812790] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.822539] env[65107]: DEBUG oslo_vmware.api [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103222, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.849169] env[65107]: ERROR nova.scheduler.client.report [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [req-82137faa-3d31-48c5-b069-4cccc795cfde] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-82137faa-3d31-48c5-b069-4cccc795cfde"}]} [ 980.872662] env[65107]: DEBUG nova.scheduler.client.report [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 980.887790] env[65107]: DEBUG nova.scheduler.client.report [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 980.888033] env[65107]: DEBUG nova.compute.provider_tree [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.902303] env[65107]: DEBUG nova.scheduler.client.report [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 980.927082] env[65107]: DEBUG nova.scheduler.client.report [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 981.034344] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103221, 'name': Rename_Task, 'duration_secs': 0.184906} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.034684] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.034951] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b58d7058-e326-4fe9-83d3-fb1c8bf21888 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.045346] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 981.045346] env[65107]: value = "task-5103223" [ 981.045346] env[65107]: _type = "Task" [ 981.045346] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.057755] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103223, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.118367] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.118729] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.119386] env[65107]: DEBUG nova.compute.manager [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 981.120814] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f37c77-1acf-4eed-8db7-4d45d9ba0bc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.129537] env[65107]: DEBUG nova.compute.manager [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 981.131106] env[65107]: DEBUG nova.objects.instance [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'flavor' on Instance uuid cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.241441] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "refresh_cache-58345821-536a-46ad-af55-92de7552e924" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 981.241930] env[65107]: DEBUG nova.compute.manager [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 981.242143] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.243067] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5565c0-cd4e-4928-818f-4d136a7b5568 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.252626] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.252900] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d4c324a-6c34-4385-b06f-7d44be727873 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.260277] env[65107]: DEBUG oslo_vmware.api [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 981.260277] env[65107]: value = "task-5103224" [ 981.260277] env[65107]: _type = "Task" [ 981.260277] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.272972] env[65107]: DEBUG oslo_vmware.api [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.326701] env[65107]: DEBUG oslo_vmware.api [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103222, 'name': ReconfigVM_Task, 'duration_secs': 0.431393} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.329957] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Reconfigured VM instance instance-0000002b to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 981.335329] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c29b33b-0fe9-4812-b735-453cc396a6e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.357026] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a58be0-c980-09e9-a375-6438f3ef29c6/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 981.358046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7dab99-7965-4d26-b3d2-8645a5e9c753 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.362457] env[65107]: DEBUG oslo_vmware.api [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 981.362457] env[65107]: value = "task-5103225" [ 981.362457] env[65107]: _type = "Task" [ 981.362457] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.370663] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a58be0-c980-09e9-a375-6438f3ef29c6/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 981.370857] env[65107]: ERROR oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a58be0-c980-09e9-a375-6438f3ef29c6/disk-0.vmdk due to incomplete transfer. [ 981.371720] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c6652d58-3989-4a46-8d55-e2bd4165a5bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.377264] env[65107]: DEBUG oslo_vmware.api [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103225, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.380674] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a58be0-c980-09e9-a375-6438f3ef29c6/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 981.380896] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Uploaded image c38e9e3f-de87-4115-b1da-c44beabd8098 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 981.383630] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 981.386737] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-537fb268-9e1c-4fee-814d-b227faaca951 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.394140] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 981.394140] env[65107]: value = "task-5103226" [ 981.394140] env[65107]: _type = "Task" [ 981.394140] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.434601] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a5becd-4ac5-42d1-b2d8-fe7b44dd75f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.442625] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cad7a02-9c0d-4ee9-993b-5d4fb0968684 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.477754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f292fdcb-e446-4a7c-a885-24ef00a1c2be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.488039] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60bc4e3-159d-405b-8a6d-0d5995d4c1e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.510384] env[65107]: DEBUG nova.compute.provider_tree [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.559556] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103223, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.669346] env[65107]: INFO nova.compute.manager [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Rebuilding instance [ 981.712127] env[65107]: DEBUG nova.compute.manager [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 981.713085] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089fe9d9-1e8b-43fc-9195-78e664f1b32b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.772070] env[65107]: DEBUG oslo_vmware.api [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103224, 'name': PowerOffVM_Task, 'duration_secs': 0.142862} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.772324] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.772520] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.773174] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c07abd4-f1c5-4b9f-b979-3c26b4432027 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.800397] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.800627] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.800811] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleting the datastore file [datastore1] 58345821-536a-46ad-af55-92de7552e924 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.801106] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4125a91a-43a7-4fcc-b9dd-14523e0bd604 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.809025] env[65107]: DEBUG oslo_vmware.api [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 981.809025] env[65107]: value = "task-5103228" [ 981.809025] env[65107]: _type = "Task" [ 981.809025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.818768] env[65107]: DEBUG oslo_vmware.api [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103228, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.874192] env[65107]: DEBUG oslo_vmware.api [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103225, 'name': ReconfigVM_Task, 'duration_secs': 0.171441} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.875051] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992733', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'name': 'volume-aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '82e8c093-b718-4d38-9682-ba2e710e4b93', 'attached_at': '', 'detached_at': '', 'volume_id': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742', 'serial': 'aae9fa95-1cf2-433e-94d4-b3c0b1119742'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 981.906374] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103226, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.046977] env[65107]: DEBUG nova.scheduler.client.report [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 982.047194] env[65107]: DEBUG nova.compute.provider_tree [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 120 to 121 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 982.047377] env[65107]: DEBUG nova.compute.provider_tree [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 982.061276] env[65107]: DEBUG oslo_vmware.api [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103223, 'name': PowerOnVM_Task, 'duration_secs': 0.597562} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.061551] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.061748] env[65107]: INFO nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Took 7.18 seconds to spawn the instance on the hypervisor. [ 982.061972] env[65107]: DEBUG nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 982.062726] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaafb61a-38b0-4c4a-a8c5-6138a25d1082 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.139655] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.139807] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afb5bf20-27eb-424e-9ce3-9e2ecca05dda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.149487] env[65107]: DEBUG oslo_vmware.api [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 982.149487] env[65107]: value = "task-5103229" [ 982.149487] env[65107]: _type = "Task" [ 982.149487] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.158356] env[65107]: DEBUG oslo_vmware.api [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.318774] env[65107]: DEBUG oslo_vmware.api [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103228, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149807} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.319052] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.319259] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.319416] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.319611] env[65107]: INFO nova.compute.manager [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Took 1.08 seconds to destroy the instance on the hypervisor. [ 982.319834] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 982.320042] env[65107]: DEBUG nova.compute.manager [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 982.320140] env[65107]: DEBUG nova.network.neutron [-] [instance: 58345821-536a-46ad-af55-92de7552e924] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 982.320391] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 982.320907] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 982.321563] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.344267] env[65107]: DEBUG nova.network.neutron [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 982.344535] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 982.406104] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103226, 'name': Destroy_Task, 'duration_secs': 0.560023} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.406104] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Destroyed the VM [ 982.406104] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 982.406374] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-028fc709-cb01-4a1a-b51c-20a2fc5c8159 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.412972] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 982.412972] env[65107]: value = "task-5103230" [ 982.412972] env[65107]: _type = "Task" [ 982.412972] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.421279] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103230, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.426264] env[65107]: DEBUG nova.objects.instance [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'flavor' on Instance uuid 82e8c093-b718-4d38-9682-ba2e710e4b93 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.556056] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.733s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.556656] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 982.559323] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.060s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.561194] env[65107]: INFO nova.compute.claims [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.580167] env[65107]: INFO nova.compute.manager [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Took 34.66 seconds to build instance. [ 982.659877] env[65107]: DEBUG oslo_vmware.api [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.730594] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.730738] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccb5415a-0074-4b61-a1ad-074f7098adba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.738982] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 982.738982] env[65107]: value = "task-5103231" [ 982.738982] env[65107]: _type = "Task" [ 982.738982] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.748505] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103231, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.847192] env[65107]: DEBUG nova.network.neutron [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 982.923332] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103230, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.066237] env[65107]: DEBUG nova.compute.utils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 983.071600] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 983.071927] env[65107]: DEBUG nova.network.neutron [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 983.072586] env[65107]: WARNING neutronclient.v2_0.client [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.073828] env[65107]: WARNING neutronclient.v2_0.client [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.073828] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.074020] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.084856] env[65107]: DEBUG oslo_concurrency.lockutils [None req-477d86d2-bc2b-4d05-b776-5b285f1df8f6 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.577s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.162589] env[65107]: DEBUG oslo_vmware.api [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103229, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.167316] env[65107]: DEBUG nova.policy [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0480b6c310e5483c9b5eb3b9b6e972bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4daf9bbfa3e9477d83d9d8c3d1fb5a02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 983.252327] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103231, 'name': PowerOffVM_Task, 'duration_secs': 0.138668} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.252577] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.253389] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.254211] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5eb8042-fc48-4f98-8639-3ab7ab49bd73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.262369] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.262535] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47485697-b4d3-4d99-8307-ce359fd14033 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.292876] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.293138] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.293334] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Deleting the datastore file [datastore1] 13c96e2f-46de-46e0-a505-60dfec9e95ca {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.293612] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44df6f33-a8e9-43f0-8cbf-a8719592afe4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.302317] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 983.302317] env[65107]: value = "task-5103233" [ 983.302317] env[65107]: _type = "Task" [ 983.302317] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.312058] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.350573] env[65107]: INFO nova.compute.manager [-] [instance: 58345821-536a-46ad-af55-92de7552e924] Took 1.03 seconds to deallocate network for instance. [ 983.425948] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103230, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.434237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d2d95de1-ce8e-4019-9bda-3a894af3b4b7 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.280s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.481149] env[65107]: DEBUG nova.network.neutron [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Successfully created port: 51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 983.571416] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 983.662616] env[65107]: DEBUG oslo_vmware.api [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103229, 'name': PowerOffVM_Task, 'duration_secs': 1.100062} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.662616] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.662616] env[65107]: DEBUG nova.compute.manager [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 983.662914] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b2bd3b-af75-4370-923d-9af515c9e6bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.732655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.732655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.732655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "82e8c093-b718-4d38-9682-ba2e710e4b93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.732655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.732932] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.735194] env[65107]: INFO nova.compute.manager [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Terminating instance [ 983.813343] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103233, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103924} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.817181] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.817181] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.817181] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.861217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.932155] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103230, 'name': RemoveSnapshot_Task, 'duration_secs': 1.231298} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.932155] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 983.932155] env[65107]: DEBUG nova.compute.manager [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 983.932485] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5785324-9b26-43f6-927b-bf1e4ac8e193 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.978548] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a8e7d5-d0fd-40e7-bc8a-f84e688ba8fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.987168] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c0fca8-3956-4c0e-bf88-2f76885c38a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.022233] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6b5532-1e0d-44db-84b5-dd68f9ef64f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.031627] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916462bb-7970-40a2-bbc4-6caee2c977af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.047680] env[65107]: DEBUG nova.compute.provider_tree [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 984.180197] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79940caa-13bf-478a-8de2-106f62ea474c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 3.061s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.239588] env[65107]: DEBUG nova.compute.manager [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 984.239821] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.240749] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f6cbe5-080a-4acc-969e-1ac7dfe57bd8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.250839] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.251411] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8ab6abc-34bc-4180-be25-63def5d24767 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.258018] env[65107]: DEBUG oslo_vmware.api [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 984.258018] env[65107]: value = "task-5103234" [ 984.258018] env[65107]: _type = "Task" [ 984.258018] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.267287] env[65107]: DEBUG oslo_vmware.api [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.451685] env[65107]: INFO nova.compute.manager [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Shelve offloading [ 984.555679] env[65107]: INFO nova.compute.manager [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Rebuilding instance [ 984.589253] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 984.591271] env[65107]: DEBUG nova.scheduler.client.report [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Updated inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with generation 121 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 984.591499] env[65107]: DEBUG nova.compute.provider_tree [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Updating resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd generation from 121 to 122 during operation: update_inventory {{(pid=65107) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 984.591673] env[65107]: DEBUG nova.compute.provider_tree [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 984.618504] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 984.618754] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 984.618908] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 984.619138] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 984.619301] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 984.619446] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 984.619653] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.619810] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 984.619976] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 984.620149] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 984.620320] env[65107]: DEBUG nova.virt.hardware [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 984.621463] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b55a31-3147-449f-94c8-671c41f14025 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.628030] env[65107]: DEBUG nova.compute.manager [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 984.628030] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4db0d74-2e5a-4e83-9752-39b7dcc53512 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.636866] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9755a866-d910-4751-82a8-67e632287a98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.770269] env[65107]: DEBUG oslo_vmware.api [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103234, 'name': PowerOffVM_Task, 'duration_secs': 0.240517} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.770269] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.770269] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.770269] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-590c1067-f676-44fb-8a6d-b33a89f6ca17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.839933] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.839933] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.840242] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleting the datastore file [datastore1] 82e8c093-b718-4d38-9682-ba2e710e4b93 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.840565] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6370c3cb-12bc-4ada-a127-f89738fd5307 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.848179] env[65107]: DEBUG oslo_vmware.api [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 984.848179] env[65107]: value = "task-5103236" [ 984.848179] env[65107]: _type = "Task" [ 984.848179] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.859506] env[65107]: DEBUG oslo_vmware.api [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103236, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.861240] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 984.861686] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 984.861686] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 984.861851] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 984.862895] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 984.862895] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 984.862895] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.862895] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 984.863308] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 984.863308] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 984.863542] env[65107]: DEBUG nova.virt.hardware [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 984.864240] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0600e943-02b6-4c22-87c9-6a5c1e1217a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.873618] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def6a7a9-c13d-4428-bc19-f282016e8642 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.889524] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.895159] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 984.895447] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.895664] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-156834a1-86d6-4842-bc25-fc053f6af6cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.910725] env[65107]: DEBUG nova.compute.manager [req-9843c610-2163-41c3-93bd-a07d017de549 req-0c18754d-59e2-4085-b65f-8306010a75ac service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Received event network-vif-plugged-51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 984.910955] env[65107]: DEBUG oslo_concurrency.lockutils [req-9843c610-2163-41c3-93bd-a07d017de549 req-0c18754d-59e2-4085-b65f-8306010a75ac service nova] Acquiring lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.911183] env[65107]: DEBUG oslo_concurrency.lockutils [req-9843c610-2163-41c3-93bd-a07d017de549 req-0c18754d-59e2-4085-b65f-8306010a75ac service nova] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.911350] env[65107]: DEBUG oslo_concurrency.lockutils [req-9843c610-2163-41c3-93bd-a07d017de549 req-0c18754d-59e2-4085-b65f-8306010a75ac service nova] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.911517] env[65107]: DEBUG nova.compute.manager [req-9843c610-2163-41c3-93bd-a07d017de549 req-0c18754d-59e2-4085-b65f-8306010a75ac service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] No waiting events found dispatching network-vif-plugged-51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 984.911676] env[65107]: WARNING nova.compute.manager [req-9843c610-2163-41c3-93bd-a07d017de549 req-0c18754d-59e2-4085-b65f-8306010a75ac service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Received unexpected event network-vif-plugged-51b0e4fe-384d-4ad6-8341-4525d8b55ffa for instance with vm_state building and task_state spawning. [ 984.919942] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.919942] env[65107]: value = "task-5103237" [ 984.919942] env[65107]: _type = "Task" [ 984.919942] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.928483] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103237, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.934431] env[65107]: DEBUG nova.objects.instance [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'flavor' on Instance uuid cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.955475] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.955897] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad4188f4-1282-4253-a13b-bd2c9aa3b1be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.964266] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 984.964266] env[65107]: value = "task-5103238" [ 984.964266] env[65107]: _type = "Task" [ 984.964266] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.973489] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.000287] env[65107]: DEBUG nova.network.neutron [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Successfully updated port: 51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 985.099036] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.099706] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 985.103138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.862s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 985.103397] env[65107]: DEBUG nova.objects.instance [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lazy-loading 'resources' on Instance uuid 179e613b-e5a2-4fbc-8fa6-b72769425ff3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.358958] env[65107]: DEBUG oslo_vmware.api [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103236, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186135} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.359403] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.359481] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.359658] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.359828] env[65107]: INFO nova.compute.manager [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Took 1.12 seconds to destroy the instance on the hypervisor. [ 985.360094] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 985.360296] env[65107]: DEBUG nova.compute.manager [-] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 985.360443] env[65107]: DEBUG nova.network.neutron [-] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 985.360709] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.361258] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.361558] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.399582] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.431895] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103237, 'name': CreateVM_Task, 'duration_secs': 0.301942} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.432108] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.432595] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.432809] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.433202] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 985.433478] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49b5febd-927f-49f4-bd61-6e31ca52a5a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.439817] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 985.439817] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b09305-b279-37b8-a8ae-11d1de6f60e6" [ 985.439817] env[65107]: _type = "Task" [ 985.439817] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.440290] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.440490] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.440610] env[65107]: DEBUG nova.network.neutron [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 985.440777] env[65107]: DEBUG nova.objects.instance [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'info_cache' on Instance uuid cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.451462] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b09305-b279-37b8-a8ae-11d1de6f60e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.476274] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 985.476512] env[65107]: DEBUG nova.compute.manager [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 985.477333] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a98cdc1-0ef4-493d-bc47-a7ad27a82c5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.485470] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.485628] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.485714] env[65107]: DEBUG nova.network.neutron [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 985.503235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "refresh_cache-6d5eb77c-c6e8-4ef1-a945-b3485aedd488" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.503480] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired lock "refresh_cache-6d5eb77c-c6e8-4ef1-a945-b3485aedd488" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.503582] env[65107]: DEBUG nova.network.neutron [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 985.607326] env[65107]: DEBUG nova.compute.utils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 985.612044] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 985.612251] env[65107]: DEBUG nova.network.neutron [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 985.612594] env[65107]: WARNING neutronclient.v2_0.client [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.613796] env[65107]: WARNING neutronclient.v2_0.client [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.613796] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.613796] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.652926] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.653434] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff81a0bb-f8f9-4365-b742-9a6dc588aacb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.663317] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 985.663317] env[65107]: value = "task-5103239" [ 985.663317] env[65107]: _type = "Task" [ 985.663317] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.675370] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.687928] env[65107]: DEBUG nova.policy [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d55cff40f12420b8320448083ca8f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7784762e36ae4c6fb30012ec801c3b88', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 985.947307] env[65107]: DEBUG nova.objects.base [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 985.955813] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b09305-b279-37b8-a8ae-11d1de6f60e6, 'name': SearchDatastore_Task, 'duration_secs': 0.011552} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.958697] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 985.958950] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.959200] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.959376] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.959517] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.963427] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eaf8fed9-1e1c-4ede-b439-eeefbdbdaad9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.979682] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.979682] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.979682] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d551d10-725a-4803-a603-1d47b8aa8e66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.984424] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 985.984424] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529c17f5-5068-6d83-e05f-7fa59da341dd" [ 985.984424] env[65107]: _type = "Task" [ 985.984424] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.991651] env[65107]: WARNING neutronclient.v2_0.client [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.991651] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.991651] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.007767] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.008404] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.015466] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529c17f5-5068-6d83-e05f-7fa59da341dd, 'name': SearchDatastore_Task, 'duration_secs': 0.01349} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.019631] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-731ca8ec-d239-4a77-bfc2-f5e18a2eb49a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.027916] env[65107]: DEBUG nova.compute.manager [req-869b3f83-8412-413d-a314-64a9726b1624 req-0242ef70-4968-430b-a3fb-7f92d119254f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Received event network-vif-deleted-8622bbd6-6068-43cb-9da2-e996d2202ef2 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 986.027916] env[65107]: INFO nova.compute.manager [req-869b3f83-8412-413d-a314-64a9726b1624 req-0242ef70-4968-430b-a3fb-7f92d119254f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Neutron deleted interface 8622bbd6-6068-43cb-9da2-e996d2202ef2; detaching it from the instance and deleting it from the info cache [ 986.027916] env[65107]: DEBUG nova.network.neutron [req-869b3f83-8412-413d-a314-64a9726b1624 req-0242ef70-4968-430b-a3fb-7f92d119254f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.029539] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 986.029539] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a28e3-776a-16ca-b69d-bf2f6ff58abf" [ 986.029539] env[65107]: _type = "Task" [ 986.029539] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.040666] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529a28e3-776a-16ca-b69d-bf2f6ff58abf, 'name': SearchDatastore_Task, 'duration_secs': 0.01197} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.043882] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 986.043882] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.043882] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9eeb136-2e93-45bf-bf96-b9d09354581f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.046243] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26148926-f392-40a0-b70e-b1a41358e20f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.060553] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e079ae28-e410-418a-8544-67e3f4cdacfe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.062678] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 986.062678] env[65107]: value = "task-5103240" [ 986.062678] env[65107]: _type = "Task" [ 986.062678] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.097078] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a8ed14-f090-4643-8b32-c8b6a708e6a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.102748] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103240, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.108870] env[65107]: DEBUG nova.network.neutron [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 986.113416] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d8a1c2-b1a4-4a06-947c-702cdabb29e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.119632] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 986.138030] env[65107]: DEBUG nova.compute.provider_tree [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.159170] env[65107]: DEBUG nova.network.neutron [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Successfully created port: 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 986.173226] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103239, 'name': PowerOffVM_Task, 'duration_secs': 0.197311} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.173594] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.173872] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 986.174757] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c398260-3ad7-4bd2-a776-4eed59a2fa50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.185742] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.187331] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.187690] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.194248] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37e4d567-7c0a-4d32-8b83-4a16f69c1f1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.214059] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.214459] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.265869] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.266226] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.266468] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.266880] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99ec8228-29eb-4c03-8678-ee8fb2b7e9be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.282349] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 986.282349] env[65107]: value = "task-5103242" [ 986.282349] env[65107]: _type = "Task" [ 986.282349] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.291831] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.301805] env[65107]: WARNING neutronclient.v2_0.client [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.302569] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.302912] env[65107]: WARNING openstack [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.332389] env[65107]: WARNING neutronclient.v2_0.client [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.333079] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.333455] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.377016] env[65107]: DEBUG nova.network.neutron [-] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.451629] env[65107]: WARNING neutronclient.v2_0.client [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.452921] env[65107]: WARNING openstack [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.452921] env[65107]: WARNING openstack [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.467151] env[65107]: DEBUG nova.network.neutron [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Updating instance_info_cache with network_info: [{"id": "51b0e4fe-384d-4ad6-8341-4525d8b55ffa", "address": "fa:16:3e:42:69:4d", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.18", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b0e4fe-38", "ovs_interfaceid": "51b0e4fe-384d-4ad6-8341-4525d8b55ffa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.531276] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c0ac995-7d11-4963-8aa6-97026aba4cd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.544877] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43e63e2-9643-4379-a203-7be77ebb1f10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.573940] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103240, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.587273] env[65107]: DEBUG nova.compute.manager [req-869b3f83-8412-413d-a314-64a9726b1624 req-0242ef70-4968-430b-a3fb-7f92d119254f service nova] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Detach interface failed, port_id=8622bbd6-6068-43cb-9da2-e996d2202ef2, reason: Instance 82e8c093-b718-4d38-9682-ba2e710e4b93 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 986.594438] env[65107]: DEBUG nova.network.neutron [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.638386] env[65107]: DEBUG nova.scheduler.client.report [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 986.794499] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382222} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.794924] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.795141] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.795324] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.806175] env[65107]: WARNING openstack [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.806175] env[65107]: WARNING openstack [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.882822] env[65107]: INFO nova.compute.manager [-] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Took 1.52 seconds to deallocate network for instance. [ 986.968370] env[65107]: WARNING neutronclient.v2_0.client [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.969052] env[65107]: WARNING openstack [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.969467] env[65107]: WARNING openstack [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.977235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Releasing lock "refresh_cache-6d5eb77c-c6e8-4ef1-a945-b3485aedd488" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 986.977579] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Instance network_info: |[{"id": "51b0e4fe-384d-4ad6-8341-4525d8b55ffa", "address": "fa:16:3e:42:69:4d", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.18", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b0e4fe-38", "ovs_interfaceid": "51b0e4fe-384d-4ad6-8341-4525d8b55ffa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 986.978289] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:69:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30c39e9a-a798-4f25-a48c-91f786ba332c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51b0e4fe-384d-4ad6-8341-4525d8b55ffa', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.986015] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 986.986252] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.986481] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c5946e0-fa0d-4fed-ba78-e32145c85469 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.009769] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.009769] env[65107]: value = "task-5103243" [ 987.009769] env[65107]: _type = "Task" [ 987.009769] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.018560] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103243, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.077469] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103240, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542931} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.077620] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.077731] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.078134] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a18bb15-1ae2-4a30-be86-762d3e45ca6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.088822] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 987.088822] env[65107]: value = "task-5103244" [ 987.088822] env[65107]: _type = "Task" [ 987.088822] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.099314] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 987.099734] env[65107]: WARNING neutronclient.v2_0.client [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.100524] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.100952] env[65107]: WARNING openstack [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.106324] env[65107]: WARNING neutronclient.v2_0.client [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.108499] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103244, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.131843] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 987.144172] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.147427] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.199s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.149073] env[65107]: INFO nova.compute.claims [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.164519] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 987.164838] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 987.165066] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 987.165299] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 987.165498] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 987.166035] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 987.166035] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.166218] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 987.166350] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 987.166550] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 987.166759] env[65107]: DEBUG nova.virt.hardware [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 987.167843] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb497de-0950-4f2f-ae3a-2fdad5b6937d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.176787] env[65107]: INFO nova.scheduler.client.report [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Deleted allocations for instance 179e613b-e5a2-4fbc-8fa6-b72769425ff3 [ 987.179288] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0907e5f9-2a07-472b-b9fd-610e67a91784 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.188192] env[65107]: DEBUG nova.network.neutron [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 987.266775] env[65107]: DEBUG nova.compute.manager [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Received event network-changed-51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 987.267191] env[65107]: DEBUG nova.compute.manager [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Refreshing instance network info cache due to event network-changed-51b0e4fe-384d-4ad6-8341-4525d8b55ffa. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 987.267510] env[65107]: DEBUG oslo_concurrency.lockutils [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Acquiring lock "refresh_cache-6d5eb77c-c6e8-4ef1-a945-b3485aedd488" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.267696] env[65107]: DEBUG oslo_concurrency.lockutils [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Acquired lock "refresh_cache-6d5eb77c-c6e8-4ef1-a945-b3485aedd488" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 987.267924] env[65107]: DEBUG nova.network.neutron [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Refreshing network info cache for port 51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 987.394138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.521187] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103243, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.587358] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.588380] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85ccbb7-81ad-4dd3-9cc8-628d9cbc6c8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.603936] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.604657] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103244, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071788} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.604925] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d0b21e2-5950-4b8f-b606-8a67f35fdf5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.606595] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.607453] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9bb1d5-9099-4060-8353-8d65d75f734b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.628371] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.628688] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4c9b786-e2cf-4283-a26f-511a1493f5dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.651092] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 987.651092] env[65107]: value = "task-5103246" [ 987.651092] env[65107]: _type = "Task" [ 987.651092] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.662202] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103246, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.676579] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.676981] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.677165] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleting the datastore file [datastore2] d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.677898] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-142a8577-05bb-4609-8dde-79697c3bf7f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.687206] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 987.687206] env[65107]: value = "task-5103247" [ 987.687206] env[65107]: _type = "Task" [ 987.687206] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.695469] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 987.697544] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3e6d79b-391e-4ea1-8446-9de45226d505 tempest-SecurityGroupsTestJSON-825765339 tempest-SecurityGroupsTestJSON-825765339-project-member] Lock "179e613b-e5a2-4fbc-8fa6-b72769425ff3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.319s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.704830] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.770505] env[65107]: WARNING neutronclient.v2_0.client [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.773184] env[65107]: WARNING openstack [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.773184] env[65107]: WARNING openstack [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.800676] env[65107]: DEBUG nova.network.neutron [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Successfully updated port: 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 987.851037] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 987.851037] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 987.851037] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 987.851466] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 987.851466] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 987.851466] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 987.851589] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.851736] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 987.851909] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 987.852565] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 987.852565] env[65107]: DEBUG nova.virt.hardware [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 987.853635] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425190f4-927f-4aa1-8db7-403f3fc012a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.863941] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d53cce6-77a8-4f44-9dde-01ad6ee2bb8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.880319] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:8b:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a34ba705-83e3-4da1-ad09-b23e6fc8d54e', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.890139] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 987.895350] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.895350] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1137ca8b-8410-4080-861c-af5428c2d000 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.922536] env[65107]: WARNING openstack [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.923772] env[65107]: WARNING openstack [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.940388] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.940388] env[65107]: value = "task-5103248" [ 987.940388] env[65107]: _type = "Task" [ 987.940388] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.953603] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103248, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.993241] env[65107]: WARNING neutronclient.v2_0.client [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.993904] env[65107]: WARNING openstack [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.995064] env[65107]: WARNING openstack [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.022186] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103243, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.089773] env[65107]: DEBUG nova.network.neutron [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Updated VIF entry in instance network info cache for port 51b0e4fe-384d-4ad6-8341-4525d8b55ffa. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 988.090188] env[65107]: DEBUG nova.network.neutron [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Updating instance_info_cache with network_info: [{"id": "51b0e4fe-384d-4ad6-8341-4525d8b55ffa", "address": "fa:16:3e:42:69:4d", "network": {"id": "3ad664c5-74aa-4c37-95db-5c8b70367282", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.18", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5b452d0d7ba0461d97515e1de6eeeb7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30c39e9a-a798-4f25-a48c-91f786ba332c", "external-id": "nsx-vlan-transportzone-438", "segmentation_id": 438, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b0e4fe-38", "ovs_interfaceid": "51b0e4fe-384d-4ad6-8341-4525d8b55ffa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 988.151870] env[65107]: DEBUG nova.compute.manager [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Received event network-vif-plugged-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 988.151870] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Acquiring lock "3df65e64-0c45-4707-960e-8f1767e2d011-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.152069] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Lock "3df65e64-0c45-4707-960e-8f1767e2d011-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.152314] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Lock "3df65e64-0c45-4707-960e-8f1767e2d011-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.152577] env[65107]: DEBUG nova.compute.manager [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] No waiting events found dispatching network-vif-plugged-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 988.152960] env[65107]: WARNING nova.compute.manager [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Received unexpected event network-vif-plugged-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d for instance with vm_state building and task_state spawning. [ 988.153081] env[65107]: DEBUG nova.compute.manager [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Received event network-changed-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 988.153258] env[65107]: DEBUG nova.compute.manager [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Refreshing instance network info cache due to event network-changed-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 988.153360] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Acquiring lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.153542] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Acquired lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.153669] env[65107]: DEBUG nova.network.neutron [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Refreshing network info cache for port 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 988.169306] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103246, 'name': ReconfigVM_Task, 'duration_secs': 0.326945} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.169624] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 13c96e2f-46de-46e0-a505-60dfec9e95ca/13c96e2f-46de-46e0-a505-60dfec9e95ca.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.170306] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a3d1fa0-e0b0-451c-82f7-638574c36dc9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.177538] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 988.177538] env[65107]: value = "task-5103249" [ 988.177538] env[65107]: _type = "Task" [ 988.177538] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.189883] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103249, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.204172] env[65107]: DEBUG oslo_vmware.api [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177753} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.207647] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.207936] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 988.208313] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 988.232113] env[65107]: INFO nova.scheduler.client.report [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted allocations for instance d2e7eee7-c9b2-49aa-92f9-1efd7c901509 [ 988.279485] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.283248] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.303598] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.452617] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103248, 'name': CreateVM_Task, 'duration_secs': 0.379359} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.452617] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.452617] env[65107]: WARNING neutronclient.v2_0.client [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 988.453071] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.453259] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.453576] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 988.454843] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a61384e-fe3b-4049-9d20-f2adc1501cd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.464855] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 988.464855] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c4085-44ac-3f0d-e28a-17028164e4f4" [ 988.464855] env[65107]: _type = "Task" [ 988.464855] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.481637] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c4085-44ac-3f0d-e28a-17028164e4f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.521890] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103243, 'name': CreateVM_Task, 'duration_secs': 1.433875} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.522102] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.522605] env[65107]: WARNING neutronclient.v2_0.client [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.522957] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.523127] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.523424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 988.523688] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d9f42f7-42a4-401a-8c4f-28f4e6b801bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.529198] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 988.529198] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a31735-e6ac-5b98-3950-630230dd650d" [ 988.529198] env[65107]: _type = "Task" [ 988.529198] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.530844] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de2d432-97a2-4c1a-ab6f-4476f348c078 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.545766] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ebc5c6-5ab6-4755-8996-e0572b2fafea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.549197] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a31735-e6ac-5b98-3950-630230dd650d, 'name': SearchDatastore_Task, 'duration_secs': 0.011452} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.549531] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.549775] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.550032] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.550167] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.550342] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.551017] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69c965cd-c2f8-4b0a-b7ba-bda0754d6366 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.582659] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a56adba-2c8f-430f-a97e-ecbd7020594c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.585546] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.585717] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.586497] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa5cd615-be45-4958-aa13-28391b466f0b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.592638] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 988.592638] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527e2993-8d1e-ccc7-89aa-eeac4f194e21" [ 988.592638] env[65107]: _type = "Task" [ 988.592638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.597485] env[65107]: DEBUG oslo_concurrency.lockutils [req-63d3aab4-0829-4e52-b597-f99d8fb41054 req-7be39956-fa45-4dcd-946b-0de7c6ca3c60 service nova] Releasing lock "refresh_cache-6d5eb77c-c6e8-4ef1-a945-b3485aedd488" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.601655] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a3cb1e-394a-4d41-8c40-cd6708ae0d43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.611327] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527e2993-8d1e-ccc7-89aa-eeac4f194e21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.619870] env[65107]: DEBUG nova.compute.provider_tree [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.656820] env[65107]: WARNING neutronclient.v2_0.client [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.658031] env[65107]: WARNING openstack [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.658031] env[65107]: WARNING openstack [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.688762] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103249, 'name': Rename_Task, 'duration_secs': 0.147721} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.689081] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.689407] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3cced4f9-995b-4189-9bf4-1cbebd0a8178 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.696970] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 988.696970] env[65107]: value = "task-5103250" [ 988.696970] env[65107]: _type = "Task" [ 988.696970] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.707266] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.707614] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.708564] env[65107]: DEBUG nova.network.neutron [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 988.710329] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a06f61c-0275-40e6-ac98-253e44e787e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.718532] env[65107]: DEBUG oslo_vmware.api [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 988.718532] env[65107]: value = "task-5103251" [ 988.718532] env[65107]: _type = "Task" [ 988.718532] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.727880] env[65107]: DEBUG oslo_vmware.api [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103251, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.737824] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.793304] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.793504] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.793785] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.793835] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.793984] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.794156] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.794313] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 988.794644] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.840804] env[65107]: DEBUG nova.network.neutron [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 988.982281] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521c4085-44ac-3f0d-e28a-17028164e4f4, 'name': SearchDatastore_Task, 'duration_secs': 0.017174} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.982641] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.982911] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.983201] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.983367] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 988.983554] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.983902] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9edf989-07fa-4219-89de-025fbbb4f723 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.994983] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.995270] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.995950] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17cd40b4-8474-4e6e-9894-251df988bb78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.002944] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 989.002944] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5220b8a2-2069-8ea7-1be4-c178b21f04f2" [ 989.002944] env[65107]: _type = "Task" [ 989.002944] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.013273] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5220b8a2-2069-8ea7-1be4-c178b21f04f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.110047] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527e2993-8d1e-ccc7-89aa-eeac4f194e21, 'name': SearchDatastore_Task, 'duration_secs': 0.030813} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.110857] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b71e1e2b-1177-43da-a3ee-32b3819c0f4d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.117087] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 989.117087] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a4ea1e-46be-c4f4-0c7c-08c5798b968e" [ 989.117087] env[65107]: _type = "Task" [ 989.117087] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.124132] env[65107]: DEBUG nova.scheduler.client.report [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.139043] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a4ea1e-46be-c4f4-0c7c-08c5798b968e, 'name': SearchDatastore_Task, 'duration_secs': 0.012904} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.139508] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.140086] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 6d5eb77c-c6e8-4ef1-a945-b3485aedd488/6d5eb77c-c6e8-4ef1-a945-b3485aedd488.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.140371] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-636b68a5-495d-4347-8367-9b8c745b2334 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.149763] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 989.149763] env[65107]: value = "task-5103252" [ 989.149763] env[65107]: _type = "Task" [ 989.149763] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.160165] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.209729] env[65107]: DEBUG oslo_vmware.api [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103250, 'name': PowerOnVM_Task, 'duration_secs': 0.50143} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.209938] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.210155] env[65107]: DEBUG nova.compute.manager [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 989.210929] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc489720-9e0e-41d1-9d10-2fed86229ec4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.229571] env[65107]: DEBUG oslo_vmware.api [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103251, 'name': PowerOnVM_Task, 'duration_secs': 0.415381} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.229887] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.232648] env[65107]: DEBUG nova.compute.manager [None req-d72196a5-cf0d-4fd6-b2d5-14d73d2121ab tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 989.232648] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed8fcb5-2a8e-4685-ba21-dbb04eb65c54 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.297544] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.344569] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb61accb-d391-4d0e-ac0d-bec8f5a4ff6a req-90c893eb-b3f9-454c-9484-da48293ab1b2 service nova] Releasing lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.344569] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.344569] env[65107]: DEBUG nova.network.neutron [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 989.378516] env[65107]: DEBUG nova.compute.manager [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-vif-unplugged-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 989.379463] env[65107]: DEBUG oslo_concurrency.lockutils [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.379706] env[65107]: DEBUG oslo_concurrency.lockutils [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.379894] env[65107]: DEBUG oslo_concurrency.lockutils [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 989.380094] env[65107]: DEBUG nova.compute.manager [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] No waiting events found dispatching network-vif-unplugged-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 989.380271] env[65107]: WARNING nova.compute.manager [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received unexpected event network-vif-unplugged-b6bba8b1-799c-4c2b-bd94-8726094948ee for instance with vm_state shelved_offloaded and task_state None. [ 989.380439] env[65107]: DEBUG nova.compute.manager [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 989.380594] env[65107]: DEBUG nova.compute.manager [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing instance network info cache due to event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 989.380779] env[65107]: DEBUG oslo_concurrency.lockutils [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.380912] env[65107]: DEBUG oslo_concurrency.lockutils [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.381085] env[65107]: DEBUG nova.network.neutron [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 989.520778] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5220b8a2-2069-8ea7-1be4-c178b21f04f2, 'name': SearchDatastore_Task, 'duration_secs': 0.010904} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.521768] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edafba6b-635d-42bb-81b3-fcea80f49667 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.530107] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 989.530107] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526c9a23-1474-f729-bf6d-12ef7c4584bb" [ 989.530107] env[65107]: _type = "Task" [ 989.530107] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.542867] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526c9a23-1474-f729-bf6d-12ef7c4584bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.633245] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 989.633907] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 989.640858] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.826s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.640858] env[65107]: DEBUG nova.objects.instance [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'resources' on Instance uuid 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.662233] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103252, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.734771] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.847084] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.847517] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.888022] env[65107]: WARNING neutronclient.v2_0.client [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 989.888022] env[65107]: WARNING openstack [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 989.888022] env[65107]: WARNING openstack [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 989.902041] env[65107]: DEBUG nova.network.neutron [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 990.043695] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526c9a23-1474-f729-bf6d-12ef7c4584bb, 'name': SearchDatastore_Task, 'duration_secs': 0.037629} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.045216] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.045570] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.052113] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.052511] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.053374] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50e7c73c-92f0-40cc-b231-db6354d6dc56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.062092] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 990.062092] env[65107]: value = "task-5103253" [ 990.062092] env[65107]: _type = "Task" [ 990.062092] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.077753] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.147866] env[65107]: DEBUG nova.compute.utils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 990.149978] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 990.150302] env[65107]: DEBUG nova.network.neutron [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 990.150715] env[65107]: WARNING neutronclient.v2_0.client [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.151067] env[65107]: WARNING neutronclient.v2_0.client [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.151692] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.152055] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.173085] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103252, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.306284] env[65107]: DEBUG nova.policy [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e10bfe6eccff4832b786e338d4225e41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df6a18c1b6724ff2a28b9b1800c9db97', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 990.354923] env[65107]: WARNING openstack [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.354923] env[65107]: WARNING openstack [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.389923] env[65107]: WARNING neutronclient.v2_0.client [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.390746] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.391964] env[65107]: WARNING openstack [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.416197] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "13c96e2f-46de-46e0-a505-60dfec9e95ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.416447] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "13c96e2f-46de-46e0-a505-60dfec9e95ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.416743] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "13c96e2f-46de-46e0-a505-60dfec9e95ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.416958] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "13c96e2f-46de-46e0-a505-60dfec9e95ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.417459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "13c96e2f-46de-46e0-a505-60dfec9e95ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.421743] env[65107]: INFO nova.compute.manager [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Terminating instance [ 990.551563] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c6b963-f2a7-4524-9ad0-af091b8365d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.561385] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6a7725-a3f7-4773-b6db-8452f37ca62f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.575589] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.600980] env[65107]: WARNING neutronclient.v2_0.client [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.601889] env[65107]: WARNING openstack [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.601967] env[65107]: WARNING openstack [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.609805] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1d9bae-8dfc-4713-9c4d-b4853e7aa72f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.621074] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a39f062-418a-404c-8308-8eed3b3ab9ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.636045] env[65107]: DEBUG nova.compute.provider_tree [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.644807] env[65107]: DEBUG nova.network.neutron [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updating instance_info_cache with network_info: [{"id": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "address": "fa:16:3e:4e:cd:29", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfc6b6-0a", "ovs_interfaceid": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 990.662556] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 990.674443] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103252, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.715058] env[65107]: DEBUG nova.network.neutron [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Successfully created port: 8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 990.884674] env[65107]: DEBUG nova.network.neutron [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updated VIF entry in instance network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 990.885441] env[65107]: DEBUG nova.network.neutron [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 990.926394] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "refresh_cache-13c96e2f-46de-46e0-a505-60dfec9e95ca" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.926726] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquired lock "refresh_cache-13c96e2f-46de-46e0-a505-60dfec9e95ca" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.926915] env[65107]: DEBUG nova.network.neutron [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 991.048149] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 991.077016] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103253, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.140136] env[65107]: DEBUG nova.scheduler.client.report [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 991.150239] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.150621] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Instance network_info: |[{"id": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "address": "fa:16:3e:4e:cd:29", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfc6b6-0a", "ovs_interfaceid": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 991.151103] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:cd:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.162448] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating folder: Project (7784762e36ae4c6fb30012ec801c3b88). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 991.162448] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3581f4a-48ec-4981-b375-8ce3c9b5baf5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.176365] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created folder: Project (7784762e36ae4c6fb30012ec801c3b88) in parent group-v992574. [ 991.176638] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating folder: Instances. Parent ref: group-v992821. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 991.177250] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-162400ee-01ad-404a-be93-2632510011a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.183653] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103252, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.529889} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.184352] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 6d5eb77c-c6e8-4ef1-a945-b3485aedd488/6d5eb77c-c6e8-4ef1-a945-b3485aedd488.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.184586] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.184822] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbfd94a8-a74d-4ad5-b0d5-c80e0e9f4504 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.195220] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 991.195220] env[65107]: value = "task-5103256" [ 991.195220] env[65107]: _type = "Task" [ 991.195220] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.195220] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created folder: Instances in parent group-v992821. [ 991.195220] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 991.198727] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.199031] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d663d7a-8acb-4b99-bd12-00f66f0721b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.222258] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103256, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.223744] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.223744] env[65107]: value = "task-5103257" [ 991.223744] env[65107]: _type = "Task" [ 991.223744] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.232775] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103257, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.389037] env[65107]: DEBUG oslo_concurrency.lockutils [req-22c02d7f-86a1-4451-96e0-0ab58b8d8446 req-c1392de7-fbe5-46c7-8eb4-44151045c141 service nova] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.429624] env[65107]: WARNING neutronclient.v2_0.client [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.430450] env[65107]: WARNING openstack [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.431042] env[65107]: WARNING openstack [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.456286] env[65107]: DEBUG nova.network.neutron [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 991.549777] env[65107]: DEBUG nova.network.neutron [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 991.577307] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103253, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.080154} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.578824] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.578824] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.578824] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc038400-b2f6-4b4f-919b-30f4aacd1e43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.585909] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 991.585909] env[65107]: value = "task-5103258" [ 991.585909] env[65107]: _type = "Task" [ 991.585909] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.599185] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103258, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.647106] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.007s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 991.649783] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.515s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 991.650285] env[65107]: DEBUG nova.objects.instance [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lazy-loading 'resources' on Instance uuid d5766a03-054f-40ea-a57e-e640664ca683 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.674037] env[65107]: INFO nova.scheduler.client.report [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocations for instance 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151 [ 991.682529] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 991.714036] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103256, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.259579} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.718199] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 991.718787] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 991.718971] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 991.719188] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 991.719338] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 991.719514] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 991.719733] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.720132] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 991.720323] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 991.720561] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 991.720784] env[65107]: DEBUG nova.virt.hardware [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 991.721388] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.722325] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990ea258-2c1b-4689-9514-001f7ad4f1b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.726039] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2222d8-2262-4907-88d0-ca282c8454d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.754260] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b1644b-23c4-4dec-9d50-9e4302778e83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.769396] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 6d5eb77c-c6e8-4ef1-a945-b3485aedd488/6d5eb77c-c6e8-4ef1-a945-b3485aedd488.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.769711] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103257, 'name': CreateVM_Task, 'duration_secs': 0.396639} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.770547] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7212087-6c77-465b-b4de-904182df5076 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.785978] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.787137] env[65107]: WARNING neutronclient.v2_0.client [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.787532] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.787690] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.788057] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 991.789073] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3819ca14-f59f-4f73-b492-bd9ed03b0e39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.801747] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 991.801747] env[65107]: value = "task-5103259" [ 991.801747] env[65107]: _type = "Task" [ 991.801747] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.806740] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 991.806740] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520e60c1-093d-88fd-2e4c-2766a6e6b471" [ 991.806740] env[65107]: _type = "Task" [ 991.806740] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.814328] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103259, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.820659] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520e60c1-093d-88fd-2e4c-2766a6e6b471, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.053234] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Releasing lock "refresh_cache-13c96e2f-46de-46e0-a505-60dfec9e95ca" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.053691] env[65107]: DEBUG nova.compute.manager [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 992.053946] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.054999] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b40322-ad2b-42b8-b463-3b2a25cf1543 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.063418] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.064113] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2923e383-b582-46c8-9d7a-43ca743ba819 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.070418] env[65107]: DEBUG oslo_vmware.api [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 992.070418] env[65107]: value = "task-5103260" [ 992.070418] env[65107]: _type = "Task" [ 992.070418] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.080043] env[65107]: DEBUG oslo_vmware.api [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103260, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.095692] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103258, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077801} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.096041] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.096892] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cabad6-2636-4d6e-9f81-2c3f9b35ce67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.122183] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.122563] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b45e18e-729c-4ac0-a834-ca027d18585c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.144414] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 992.144414] env[65107]: value = "task-5103261" [ 992.144414] env[65107]: _type = "Task" [ 992.144414] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.157522] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103261, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.188265] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fc21f708-b4ce-49ef-ad38-f2c025f59679 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "23c3b100-30ac-44f9-8e2d-d8e3f2f4e151" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.993s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.301614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "a3162257-ac6b-4468-a573-3611bba69c46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 992.301923] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "a3162257-ac6b-4468-a573-3611bba69c46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 992.320451] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103259, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.329307] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520e60c1-093d-88fd-2e4c-2766a6e6b471, 'name': SearchDatastore_Task, 'duration_secs': 0.014361} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.330737] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.330980] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.331229] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.331378] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.331595] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.332661] env[65107]: DEBUG nova.compute.manager [req-fb38816f-d939-44c2-8ddd-a69ffe13cb5a req-600e1b4d-c3c0-4ec3-934c-d2d178160880 service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Received event network-vif-plugged-8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 992.332864] env[65107]: DEBUG oslo_concurrency.lockutils [req-fb38816f-d939-44c2-8ddd-a69ffe13cb5a req-600e1b4d-c3c0-4ec3-934c-d2d178160880 service nova] Acquiring lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 992.333074] env[65107]: DEBUG oslo_concurrency.lockutils [req-fb38816f-d939-44c2-8ddd-a69ffe13cb5a req-600e1b4d-c3c0-4ec3-934c-d2d178160880 service nova] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 992.334237] env[65107]: DEBUG oslo_concurrency.lockutils [req-fb38816f-d939-44c2-8ddd-a69ffe13cb5a req-600e1b4d-c3c0-4ec3-934c-d2d178160880 service nova] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.334237] env[65107]: DEBUG nova.compute.manager [req-fb38816f-d939-44c2-8ddd-a69ffe13cb5a req-600e1b4d-c3c0-4ec3-934c-d2d178160880 service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] No waiting events found dispatching network-vif-plugged-8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 992.334237] env[65107]: WARNING nova.compute.manager [req-fb38816f-d939-44c2-8ddd-a69ffe13cb5a req-600e1b4d-c3c0-4ec3-934c-d2d178160880 service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Received unexpected event network-vif-plugged-8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 for instance with vm_state building and task_state spawning. [ 992.336708] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84af390c-6d70-4ff4-8471-84fad1ad3f14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.353529] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.353529] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.353999] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c358738-0a39-49a5-8342-2d90a0501510 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.362954] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 992.362954] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a01dd4-6c8c-175b-d976-f0fd019b5c6f" [ 992.362954] env[65107]: _type = "Task" [ 992.362954] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.374990] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a01dd4-6c8c-175b-d976-f0fd019b5c6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.453349] env[65107]: DEBUG nova.network.neutron [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Successfully updated port: 8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 992.563447] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8fa8e8-c872-4d77-98e6-147b72ee5ae2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.575395] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b1c846-8627-45cd-a54a-84d31bb32468 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.584020] env[65107]: DEBUG oslo_vmware.api [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103260, 'name': PowerOffVM_Task, 'duration_secs': 0.137568} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.612851] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.612851] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.613909] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a7a06e9-2555-4107-8ee4-9ab4079b6794 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.616217] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ad447f-f64b-4086-a00a-e6bfd931440f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.625083] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dafb8275-fbaa-449d-864e-76642fd08f22 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.641312] env[65107]: DEBUG nova.compute.provider_tree [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.645535] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.645841] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.645981] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Deleting the datastore file [datastore2] 13c96e2f-46de-46e0-a505-60dfec9e95ca {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.649652] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11c92a53-7fa0-4621-b82e-86aaa89393d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.660070] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103261, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.662522] env[65107]: DEBUG oslo_vmware.api [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for the task: (returnval){ [ 992.662522] env[65107]: value = "task-5103263" [ 992.662522] env[65107]: _type = "Task" [ 992.662522] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.672529] env[65107]: DEBUG oslo_vmware.api [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103263, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.807598] env[65107]: DEBUG nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 992.817712] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103259, 'name': ReconfigVM_Task, 'duration_secs': 0.769687} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.818158] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 6d5eb77c-c6e8-4ef1-a945-b3485aedd488/6d5eb77c-c6e8-4ef1-a945-b3485aedd488.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.818830] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ffd1803-581c-438d-bc23-25080f65fd46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.826237] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 992.826237] env[65107]: value = "task-5103264" [ 992.826237] env[65107]: _type = "Task" [ 992.826237] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.836023] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103264, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.874365] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a01dd4-6c8c-175b-d976-f0fd019b5c6f, 'name': SearchDatastore_Task, 'duration_secs': 0.048087} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.875149] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a9886d5-8a24-4afe-a6af-f2cf8dc45928 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.882393] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 992.882393] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e5f762-8a85-c195-e096-d677091dcd78" [ 992.882393] env[65107]: _type = "Task" [ 992.882393] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.892476] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e5f762-8a85-c195-e096-d677091dcd78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.957261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "refresh_cache-59ff1aa6-8d0d-4885-90d3-33d623dca8bb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.957501] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquired lock "refresh_cache-59ff1aa6-8d0d-4885-90d3-33d623dca8bb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.957688] env[65107]: DEBUG nova.network.neutron [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 993.145369] env[65107]: DEBUG nova.scheduler.client.report [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 993.158216] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103261, 'name': ReconfigVM_Task, 'duration_secs': 0.668461} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.158509] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21/6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.159170] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b7ccfba-2259-4df9-bf57-8ed64ffb1f21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.170431] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 993.170431] env[65107]: value = "task-5103265" [ 993.170431] env[65107]: _type = "Task" [ 993.170431] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.176859] env[65107]: DEBUG oslo_vmware.api [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Task: {'id': task-5103263, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101912} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.177551] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.177755] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 993.177958] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 993.178177] env[65107]: INFO nova.compute.manager [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Took 1.12 seconds to destroy the instance on the hypervisor. [ 993.178474] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 993.178686] env[65107]: DEBUG nova.compute.manager [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 993.178798] env[65107]: DEBUG nova.network.neutron [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 993.179095] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.179683] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.179934] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.190686] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103265, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.206558] env[65107]: DEBUG nova.network.neutron [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 993.206828] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.337609] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103264, 'name': Rename_Task, 'duration_secs': 0.166528} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.338550] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 993.339215] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.339215] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75f34ad8-059f-44ca-81b5-0e118768e720 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.345766] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 993.345766] env[65107]: value = "task-5103266" [ 993.345766] env[65107]: _type = "Task" [ 993.345766] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.354272] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.394369] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e5f762-8a85-c195-e096-d677091dcd78, 'name': SearchDatastore_Task, 'duration_secs': 0.014118} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.394369] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 993.394498] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 3df65e64-0c45-4707-960e-8f1767e2d011/3df65e64-0c45-4707-960e-8f1767e2d011.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 993.394820] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8d0184d-a4de-4ddf-ae52-d48618284f43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.402099] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 993.402099] env[65107]: value = "task-5103267" [ 993.402099] env[65107]: _type = "Task" [ 993.402099] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.410653] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103267, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.462958] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.463309] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.511456] env[65107]: DEBUG nova.network.neutron [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 993.539231] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.539629] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.546403] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 993.546641] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.613619] env[65107]: WARNING neutronclient.v2_0.client [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.614834] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.614834] env[65107]: WARNING openstack [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.651418] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.656035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.635s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.656807] env[65107]: INFO nova.compute.claims [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.678180] env[65107]: INFO nova.scheduler.client.report [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Deleted allocations for instance d5766a03-054f-40ea-a57e-e640664ca683 [ 993.686865] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103265, 'name': Rename_Task, 'duration_secs': 0.305363} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.687849] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.688235] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95b53b62-93ad-45d2-ac31-fe1d8ff2075d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.699216] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 993.699216] env[65107]: value = "task-5103268" [ 993.699216] env[65107]: _type = "Task" [ 993.699216] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.712266] env[65107]: DEBUG nova.network.neutron [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 993.713735] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.726195] env[65107]: DEBUG nova.network.neutron [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Updating instance_info_cache with network_info: [{"id": "8f09ac6e-ef41-4d62-b1fd-acf86cfbed59", "address": "fa:16:3e:15:5e:60", "network": {"id": "a7157682-ebfb-4636-bd64-398799743539", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-542156413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df6a18c1b6724ff2a28b9b1800c9db97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f09ac6e-ef", "ovs_interfaceid": "8f09ac6e-ef41-4d62-b1fd-acf86cfbed59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 993.858913] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103266, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.915139] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103267, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.049841] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 994.191721] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0559a8d6-435e-4af3-969e-947f3f8180cd tempest-InstanceActionsNegativeTestJSON-930462260 tempest-InstanceActionsNegativeTestJSON-930462260-project-member] Lock "d5766a03-054f-40ea-a57e-e640664ca683" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.964s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.211739] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103268, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.217126] env[65107]: INFO nova.compute.manager [-] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Took 1.04 seconds to deallocate network for instance. [ 994.234063] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Releasing lock "refresh_cache-59ff1aa6-8d0d-4885-90d3-33d623dca8bb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.234063] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Instance network_info: |[{"id": "8f09ac6e-ef41-4d62-b1fd-acf86cfbed59", "address": "fa:16:3e:15:5e:60", "network": {"id": "a7157682-ebfb-4636-bd64-398799743539", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-542156413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df6a18c1b6724ff2a28b9b1800c9db97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f09ac6e-ef", "ovs_interfaceid": "8f09ac6e-ef41-4d62-b1fd-acf86cfbed59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 994.234063] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:5e:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f09ac6e-ef41-4d62-b1fd-acf86cfbed59', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.242846] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Creating folder: Project (df6a18c1b6724ff2a28b9b1800c9db97). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.243351] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1a40539-7759-4a05-a5a9-37e92d80f283 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.258028] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Created folder: Project (df6a18c1b6724ff2a28b9b1800c9db97) in parent group-v992574. [ 994.258711] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Creating folder: Instances. Parent ref: group-v992824. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 994.259556] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f499c891-d2f6-4386-8b2e-8b1f2afaae61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.271724] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Created folder: Instances in parent group-v992824. [ 994.272328] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 994.272781] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.273169] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89445e83-6055-4162-8fac-945e2af9289a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.296320] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.296320] env[65107]: value = "task-5103271" [ 994.296320] env[65107]: _type = "Task" [ 994.296320] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.307104] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103271, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.358341] env[65107]: DEBUG oslo_vmware.api [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103266, 'name': PowerOnVM_Task, 'duration_secs': 0.787101} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.358672] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.358913] env[65107]: INFO nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Took 9.77 seconds to spawn the instance on the hypervisor. [ 994.359076] env[65107]: DEBUG nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 994.360080] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2639d4a-3401-4a5c-8711-49b101aa60f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.373432] env[65107]: DEBUG nova.compute.manager [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Received event network-changed-8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 994.373608] env[65107]: DEBUG nova.compute.manager [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Refreshing instance network info cache due to event network-changed-8f09ac6e-ef41-4d62-b1fd-acf86cfbed59. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 994.373824] env[65107]: DEBUG oslo_concurrency.lockutils [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Acquiring lock "refresh_cache-59ff1aa6-8d0d-4885-90d3-33d623dca8bb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.373963] env[65107]: DEBUG oslo_concurrency.lockutils [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Acquired lock "refresh_cache-59ff1aa6-8d0d-4885-90d3-33d623dca8bb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 994.374297] env[65107]: DEBUG nova.network.neutron [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Refreshing network info cache for port 8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 994.414009] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103267, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558883} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.415191] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 3df65e64-0c45-4707-960e-8f1767e2d011/3df65e64-0c45-4707-960e-8f1767e2d011.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 994.415468] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.415759] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5ab4870-9666-4459-9014-8275908bfe8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.424374] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 994.424374] env[65107]: value = "task-5103272" [ 994.424374] env[65107]: _type = "Task" [ 994.424374] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.434069] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103272, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.575462] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.711171] env[65107]: DEBUG oslo_vmware.api [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103268, 'name': PowerOnVM_Task, 'duration_secs': 0.814265} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.711461] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.711660] env[65107]: DEBUG nova.compute.manager [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 994.712977] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e471339-570d-4a4e-8498-1ac4bacf5e46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.723035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 994.810371] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103271, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.880344] env[65107]: WARNING neutronclient.v2_0.client [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 994.880988] env[65107]: WARNING openstack [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.882038] env[65107]: WARNING openstack [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.897975] env[65107]: INFO nova.compute.manager [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Took 42.00 seconds to build instance. [ 994.933905] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103272, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108288} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.934304] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.935102] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77997c34-74c4-4e8f-a225-15b21a9627d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.962415] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 3df65e64-0c45-4707-960e-8f1767e2d011/3df65e64-0c45-4707-960e-8f1767e2d011.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.965375] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68661e9a-8351-495c-b5a0-366ae4c68c1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.989766] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 994.989766] env[65107]: value = "task-5103273" [ 994.989766] env[65107]: _type = "Task" [ 994.989766] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.002359] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103273, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.003647] env[65107]: WARNING openstack [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.004170] env[65107]: WARNING openstack [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.028957] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167d2034-e931-478c-a006-2fd25da72497 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.042653] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e669c3c4-eb7e-4db5-845a-f72079168a0c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.075199] env[65107]: WARNING neutronclient.v2_0.client [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 995.075839] env[65107]: WARNING openstack [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.076267] env[65107]: WARNING openstack [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.085098] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb5b310-3d62-4a0e-9594-510fda0b6955 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.094159] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5479c41-0720-4cc2-8020-ecbf89f9ead2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.110922] env[65107]: DEBUG nova.compute.provider_tree [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.186651] env[65107]: DEBUG nova.network.neutron [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Updated VIF entry in instance network info cache for port 8f09ac6e-ef41-4d62-b1fd-acf86cfbed59. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 995.186891] env[65107]: DEBUG nova.network.neutron [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Updating instance_info_cache with network_info: [{"id": "8f09ac6e-ef41-4d62-b1fd-acf86cfbed59", "address": "fa:16:3e:15:5e:60", "network": {"id": "a7157682-ebfb-4636-bd64-398799743539", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-542156413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df6a18c1b6724ff2a28b9b1800c9db97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f09ac6e-ef", "ovs_interfaceid": "8f09ac6e-ef41-4d62-b1fd-acf86cfbed59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 995.234772] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.310437] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103271, 'name': CreateVM_Task, 'duration_secs': 0.534586} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.310740] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.312372] env[65107]: WARNING neutronclient.v2_0.client [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 995.312953] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.313278] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.313736] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 995.314144] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac3a214d-4fd5-42e9-b162-6d849e46718c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.322260] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 995.322260] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0d5f5-4698-84c1-5f7f-d6062c5aafa1" [ 995.322260] env[65107]: _type = "Task" [ 995.322260] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.334125] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0d5f5-4698-84c1-5f7f-d6062c5aafa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.402817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b0b011b-4373-4c16-b605-2bcd3180bcd3 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.531s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.501105] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.615447] env[65107]: DEBUG nova.scheduler.client.report [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.692175] env[65107]: DEBUG oslo_concurrency.lockutils [req-9cc43ae4-7639-4156-9b30-3bf2cc0fac39 req-410dcaaf-9b67-42bd-85d7-7d2b3588d2ec service nova] Releasing lock "refresh_cache-59ff1aa6-8d0d-4885-90d3-33d623dca8bb" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.835852] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0d5f5-4698-84c1-5f7f-d6062c5aafa1, 'name': SearchDatastore_Task, 'duration_secs': 0.051304} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.836440] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.836626] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.836867] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.837024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.837524] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.837524] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14544346-3eb9-489a-825c-fd6a251c533c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.857044] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.857808] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.858515] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df6c92ee-9207-422d-b3d5-051da49c13fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.865763] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 995.865763] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fd7325-fd0f-195b-1de5-8fe9982e6f2e" [ 995.865763] env[65107]: _type = "Task" [ 995.865763] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.875551] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fd7325-fd0f-195b-1de5-8fe9982e6f2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.002451] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103273, 'name': ReconfigVM_Task, 'duration_secs': 0.744692} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.002767] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 3df65e64-0c45-4707-960e-8f1767e2d011/3df65e64-0c45-4707-960e-8f1767e2d011.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.003486] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4cabe1a2-0ac8-49b8-86b2-bb568b90a16c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.012662] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 996.012662] env[65107]: value = "task-5103274" [ 996.012662] env[65107]: _type = "Task" [ 996.012662] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.025187] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103274, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.126016] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.126810] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 996.132102] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.565s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.132102] env[65107]: DEBUG nova.objects.instance [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lazy-loading 'resources' on Instance uuid 52bd7ecb-4f53-443e-8916-3ff50dfab2b6 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.241877] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13db8b59-5525-406a-97df-320ee2f7dd28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.250126] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Suspending the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 996.250126] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-eabe0240-b835-41cd-a458-a828b4c82793 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.259377] env[65107]: DEBUG oslo_vmware.api [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] Waiting for the task: (returnval){ [ 996.259377] env[65107]: value = "task-5103275" [ 996.259377] env[65107]: _type = "Task" [ 996.259377] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.271936] env[65107]: DEBUG oslo_vmware.api [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] Task: {'id': task-5103275, 'name': SuspendVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.279554] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.279847] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.280634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.280634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.280634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.282933] env[65107]: INFO nova.compute.manager [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Terminating instance [ 996.377021] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fd7325-fd0f-195b-1de5-8fe9982e6f2e, 'name': SearchDatastore_Task, 'duration_secs': 0.014099} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.378046] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-518ff42a-5610-4e08-bbe8-4919c9174013 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.384913] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 996.384913] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52037b3a-b582-d7ce-29ee-2119c871105c" [ 996.384913] env[65107]: _type = "Task" [ 996.384913] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.394557] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52037b3a-b582-d7ce-29ee-2119c871105c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.523373] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103274, 'name': Rename_Task, 'duration_secs': 0.17851} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.523685] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 996.523942] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16b80965-c41c-44ac-8858-225431326ef7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.532107] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 996.532107] env[65107]: value = "task-5103276" [ 996.532107] env[65107]: _type = "Task" [ 996.532107] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.544860] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103276, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.634525] env[65107]: DEBUG nova.compute.utils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 996.639229] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 996.639519] env[65107]: DEBUG nova.network.neutron [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 996.640366] env[65107]: WARNING neutronclient.v2_0.client [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.640366] env[65107]: WARNING neutronclient.v2_0.client [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.640778] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.640962] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.723364] env[65107]: DEBUG nova.policy [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6590cb74e83e45bd944f23fa888a4b04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f379144b78764fe394039d87b043a946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 996.773982] env[65107]: DEBUG oslo_vmware.api [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] Task: {'id': task-5103275, 'name': SuspendVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.788327] env[65107]: DEBUG nova.compute.manager [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 996.788680] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.790764] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37966630-9524-4c6d-ba82-d7ac1b361491 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.804040] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.804528] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f88fbf3-cec5-4937-96e9-866b54424aa5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.813254] env[65107]: DEBUG oslo_vmware.api [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 996.813254] env[65107]: value = "task-5103277" [ 996.813254] env[65107]: _type = "Task" [ 996.813254] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.831162] env[65107]: DEBUG oslo_vmware.api [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.904196] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52037b3a-b582-d7ce-29ee-2119c871105c, 'name': SearchDatastore_Task, 'duration_secs': 0.024413} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.904467] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.904717] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 59ff1aa6-8d0d-4885-90d3-33d623dca8bb/59ff1aa6-8d0d-4885-90d3-33d623dca8bb.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.905135] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2aeb107-b1c6-4f29-bf89-a9c07309fc06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.919598] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 996.919598] env[65107]: value = "task-5103278" [ 996.919598] env[65107]: _type = "Task" [ 996.919598] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.928411] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.045509] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103276, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.121712] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549f515a-1cdf-429a-b36e-1a334ffe3164 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.135683] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51c5c38-dfb1-4bf6-9626-1a41fabb3073 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.140800] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 997.187673] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b5b2b2-4c92-4200-8447-8cde149f05b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.202078] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc961c60-f321-4457-82cf-3f658ae2459b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.211201] env[65107]: DEBUG nova.network.neutron [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Successfully created port: 416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 997.232557] env[65107]: DEBUG nova.compute.provider_tree [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.242506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.242834] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.275278] env[65107]: DEBUG oslo_vmware.api [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] Task: {'id': task-5103275, 'name': SuspendVM_Task, 'duration_secs': 0.801013} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.277054] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Suspended the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 997.277054] env[65107]: DEBUG nova.compute.manager [None req-8548ac53-b052-4cc5-b302-a2fa778a0b5a tempest-ServersAdminNegativeTestJSON-417004909 tempest-ServersAdminNegativeTestJSON-417004909-project-admin] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 997.277302] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc95a45-5a86-4881-b239-20ef65895fd6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.328806] env[65107]: DEBUG oslo_vmware.api [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103277, 'name': PowerOffVM_Task, 'duration_secs': 0.294129} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.330178] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.330386] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.330677] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19876290-e0f5-4732-a11f-fa5c49fcd6bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.408750] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.409156] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.409298] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore1] 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.409660] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89a00dff-5027-44b4-821d-30b8ca3716a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.424583] env[65107]: DEBUG oslo_vmware.api [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 997.424583] env[65107]: value = "task-5103280" [ 997.424583] env[65107]: _type = "Task" [ 997.424583] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.433049] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103278, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.443914] env[65107]: DEBUG oslo_vmware.api [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103280, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.543324] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103276, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.737538] env[65107]: DEBUG nova.scheduler.client.report [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 997.747697] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 997.930091] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564458} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.933390] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 59ff1aa6-8d0d-4885-90d3-33d623dca8bb/59ff1aa6-8d0d-4885-90d3-33d623dca8bb.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.933735] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.934067] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1ab6a8b-1c9a-4103-90ee-5610fc32c8da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.943273] env[65107]: DEBUG oslo_vmware.api [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103280, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.945576] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.945915] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.946170] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.946403] env[65107]: INFO nova.compute.manager [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Took 1.16 seconds to destroy the instance on the hypervisor. [ 997.946753] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 997.947280] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 997.947280] env[65107]: value = "task-5103281" [ 997.947280] env[65107]: _type = "Task" [ 997.947280] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.948222] env[65107]: DEBUG nova.compute.manager [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 997.948384] env[65107]: DEBUG nova.network.neutron [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 997.948980] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 997.949650] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.950117] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.967329] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103281, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.043216] env[65107]: DEBUG oslo_vmware.api [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103276, 'name': PowerOnVM_Task, 'duration_secs': 1.280224} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.043493] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.043696] env[65107]: INFO nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Took 10.91 seconds to spawn the instance on the hypervisor. [ 998.043873] env[65107]: DEBUG nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 998.044754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85f9029-f3d3-4c17-b69e-d8a10a54a889 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.151646] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 998.185252] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 998.185540] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 998.185706] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 998.185889] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 998.186053] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 998.186206] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 998.186416] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.186628] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 998.186819] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 998.186987] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 998.187176] env[65107]: DEBUG nova.virt.hardware [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 998.188152] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af525f4c-a70b-4590-8820-1ae39a9ed772 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.197174] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6254b809-7c9f-4a56-91d7-1198b91366d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.246532] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.116s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.249148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.459s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.249355] env[65107]: DEBUG nova.objects.instance [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: 58345821-536a-46ad-af55-92de7552e924] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 998.270718] env[65107]: INFO nova.scheduler.client.report [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted allocations for instance 52bd7ecb-4f53-443e-8916-3ff50dfab2b6 [ 998.272614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.466785] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103281, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092601} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.467084] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.467892] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a90384-d740-4acb-afd3-b363793ec543 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.490421] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 59ff1aa6-8d0d-4885-90d3-33d623dca8bb/59ff1aa6-8d0d-4885-90d3-33d623dca8bb.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.490756] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38dfb898-3d5c-4108-9a1f-c5f75baebc1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.513800] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 998.513800] env[65107]: value = "task-5103282" [ 998.513800] env[65107]: _type = "Task" [ 998.513800] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.522391] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103282, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.565424] env[65107]: INFO nova.compute.manager [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Took 45.09 seconds to build instance. [ 998.702203] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.781515] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be381737-99f2-458d-be3a-6a73ab8ff473 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "52bd7ecb-4f53-443e-8916-3ff50dfab2b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.707s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.812891] env[65107]: DEBUG nova.network.neutron [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Successfully updated port: 416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 999.024906] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103282, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.067560] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89ca3178-9275-4636-a4a2-6eb861311296 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "3df65e64-0c45-4707-960e-8f1767e2d011" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.614s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.260301] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f935d77a-a0fd-4d5f-88ca-54174d230ea7 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.261820] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.451s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.262081] env[65107]: DEBUG nova.objects.instance [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lazy-loading 'resources' on Instance uuid f4b1352b-1c55-4987-a298-69431c4e565d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.318565] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.318565] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 999.318565] env[65107]: DEBUG nova.network.neutron [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 999.383549] env[65107]: DEBUG nova.compute.manager [req-24667a74-8c96-4112-ac30-4c753be3d03e req-1d131873-525f-411f-97a8-5d7a5e391552 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Received event network-vif-plugged-416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 999.383948] env[65107]: DEBUG oslo_concurrency.lockutils [req-24667a74-8c96-4112-ac30-4c753be3d03e req-1d131873-525f-411f-97a8-5d7a5e391552 service nova] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.385038] env[65107]: DEBUG oslo_concurrency.lockutils [req-24667a74-8c96-4112-ac30-4c753be3d03e req-1d131873-525f-411f-97a8-5d7a5e391552 service nova] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.385038] env[65107]: DEBUG oslo_concurrency.lockutils [req-24667a74-8c96-4112-ac30-4c753be3d03e req-1d131873-525f-411f-97a8-5d7a5e391552 service nova] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.385038] env[65107]: DEBUG nova.compute.manager [req-24667a74-8c96-4112-ac30-4c753be3d03e req-1d131873-525f-411f-97a8-5d7a5e391552 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] No waiting events found dispatching network-vif-plugged-416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 999.385249] env[65107]: WARNING nova.compute.manager [req-24667a74-8c96-4112-ac30-4c753be3d03e req-1d131873-525f-411f-97a8-5d7a5e391552 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Received unexpected event network-vif-plugged-416a2590-6a42-4a32-a922-a9d5f95786d6 for instance with vm_state building and task_state spawning. [ 999.526791] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103282, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.568476] env[65107]: DEBUG nova.network.neutron [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 999.821124] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.821487] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.878231] env[65107]: DEBUG nova.network.neutron [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 999.900749] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.901751] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.999272] env[65107]: WARNING neutronclient.v2_0.client [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.999925] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1000.001348] env[65107]: WARNING openstack [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1000.026941] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103282, 'name': ReconfigVM_Task, 'duration_secs': 1.085961} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.027369] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 59ff1aa6-8d0d-4885-90d3-33d623dca8bb/59ff1aa6-8d0d-4885-90d3-33d623dca8bb.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.028106] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e474dff9-80bb-47d2-9c6c-a08db7e7adba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.042032] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 1000.042032] env[65107]: value = "task-5103283" [ 1000.042032] env[65107]: _type = "Task" [ 1000.042032] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.061366] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103283, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.070611] env[65107]: INFO nova.compute.manager [-] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Took 2.12 seconds to deallocate network for instance. [ 1000.115894] env[65107]: DEBUG nova.network.neutron [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1000.208415] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd6c748-65e4-497b-b5a0-c36153ce0890 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.220022] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad534296-264f-4b3e-b4be-73af1d73a04c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.258598] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c08425-3b0e-4f3f-b443-3449f9550326 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.268199] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9fd283-4a6d-451a-b6b4-e2e3a5f2c491 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.288914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "ac631507-8abf-4019-bdd5-7684dd51d1e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.290042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.291158] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "ac631507-8abf-4019-bdd5-7684dd51d1e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.291432] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.291621] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.296019] env[65107]: DEBUG nova.compute.provider_tree [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.296019] env[65107]: INFO nova.compute.manager [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Terminating instance [ 1000.555026] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103283, 'name': Rename_Task, 'duration_secs': 0.176743} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.555026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.555026] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d300ee2b-9399-4461-8d48-9604b2d8c79b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.560970] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 1000.560970] env[65107]: value = "task-5103284" [ 1000.560970] env[65107]: _type = "Task" [ 1000.560970] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.579144] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.579331] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.620292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1000.620833] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Instance network_info: |[{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1000.621495] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:c9:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd903c404-a23a-40c0-a217-96d4bb2e5b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '416a2590-6a42-4a32-a922-a9d5f95786d6', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.632456] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1000.632991] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.633488] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40ea0551-b6b7-4173-be45-8ec5d735f60b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.670542] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.670542] env[65107]: value = "task-5103285" [ 1000.670542] env[65107]: _type = "Task" [ 1000.670542] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.679429] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103285, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.798257] env[65107]: DEBUG nova.scheduler.client.report [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1000.803917] env[65107]: DEBUG nova.compute.manager [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1000.804164] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.805503] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bc1850-d4d0-4e13-958b-f9d132c5c66e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.815997] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.816383] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebc6e1b4-90f7-4536-aec8-233c42adda69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.824569] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1000.824569] env[65107]: value = "task-5103286" [ 1000.824569] env[65107]: _type = "Task" [ 1000.824569] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.834996] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.078208] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103284, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.112905] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.112905] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1001.184926] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103285, 'name': CreateVM_Task, 'duration_secs': 0.395436} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.185023] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.186431] env[65107]: WARNING neutronclient.v2_0.client [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.186431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.186431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.186431] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1001.186929] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fc07e35-fc76-4b75-837c-91e71da3129b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.193031] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1001.193031] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f3c78d-db61-3c05-a991-9736fb7136e7" [ 1001.193031] env[65107]: _type = "Task" [ 1001.193031] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.206473] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f3c78d-db61-3c05-a991-9736fb7136e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.240538] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.240767] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1001.307147] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.045s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.309959] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.449s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1001.309959] env[65107]: DEBUG nova.objects.instance [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lazy-loading 'resources' on Instance uuid 58345821-536a-46ad-af55-92de7552e924 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.336161] env[65107]: INFO nova.scheduler.client.report [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Deleted allocations for instance f4b1352b-1c55-4987-a298-69431c4e565d [ 1001.340862] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103286, 'name': PowerOffVM_Task, 'duration_secs': 0.2547} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.343433] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.343617] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.344099] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d1d8ee0-8e13-468f-8908-c46cea73319a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.424602] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.424833] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.425024] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleting the datastore file [datastore1] ac631507-8abf-4019-bdd5-7684dd51d1e9 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.426785] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81e091d3-d1b5-4564-b663-474d50095fdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.430652] env[65107]: DEBUG nova.compute.manager [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Received event network-changed-416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1001.431141] env[65107]: DEBUG nova.compute.manager [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Refreshing instance network info cache due to event network-changed-416a2590-6a42-4a32-a922-a9d5f95786d6. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1001.431141] env[65107]: DEBUG oslo_concurrency.lockutils [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.431141] env[65107]: DEBUG oslo_concurrency.lockutils [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.431244] env[65107]: DEBUG nova.network.neutron [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Refreshing network info cache for port 416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1001.440727] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1001.440727] env[65107]: value = "task-5103288" [ 1001.440727] env[65107]: _type = "Task" [ 1001.440727] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.452074] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.576384] env[65107]: DEBUG oslo_vmware.api [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103284, 'name': PowerOnVM_Task, 'duration_secs': 0.549855} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.576541] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.576778] env[65107]: INFO nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Took 9.90 seconds to spawn the instance on the hypervisor. [ 1001.576963] env[65107]: DEBUG nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1001.577837] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b64da55-4862-4116-9296-a9cbacb3da2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.616182] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1001.704938] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f3c78d-db61-3c05-a991-9736fb7136e7, 'name': SearchDatastore_Task, 'duration_secs': 0.017077} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.705280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.705578] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.705851] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.706069] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.706266] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.706894] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4de6e885-304f-48e2-a168-4f7c4986513a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.722202] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.722202] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.723192] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50b75c9b-6dc8-4740-9e17-8a196f312b67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.733331] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1001.733331] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52232c5a-3865-e990-28eb-747dd1350856" [ 1001.733331] env[65107]: _type = "Task" [ 1001.733331] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.747343] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1001.751032] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52232c5a-3865-e990-28eb-747dd1350856, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.851514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a78a2af5-c1e3-4eed-9c62-3af1e38e751f tempest-ServerPasswordTestJSON-1404653062 tempest-ServerPasswordTestJSON-1404653062-project-member] Lock "f4b1352b-1c55-4987-a298-69431c4e565d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.527s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.934871] env[65107]: WARNING neutronclient.v2_0.client [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.934871] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.935053] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.957582] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.108099] env[65107]: INFO nova.compute.manager [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Took 44.18 seconds to build instance. [ 1002.132049] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.132524] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.147075] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.152670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.152670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.152670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.152993] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.152993] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.157493] env[65107]: INFO nova.compute.manager [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Terminating instance [ 1002.222920] env[65107]: WARNING neutronclient.v2_0.client [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.223696] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.224068] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.249977] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52232c5a-3865-e990-28eb-747dd1350856, 'name': SearchDatastore_Task, 'duration_secs': 0.014289} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.251065] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ab2db5d-f841-4ebc-b9ec-77155f05ea90 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.257382] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66178519-5dd1-46dc-b07b-52291c22ba39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.267385] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1002.267385] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527f8d64-3d12-a2bb-34f2-bfe272d39612" [ 1002.267385] env[65107]: _type = "Task" [ 1002.267385] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.274219] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ff3644-1549-4337-b59e-519f1974b0c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.283873] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527f8d64-3d12-a2bb-34f2-bfe272d39612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.310978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.312290] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b21885b-0d5f-4097-9b97-56b5d3085e37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.321793] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e62f78-06e6-4034-bef4-a6aed524ef64 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.338151] env[65107]: DEBUG nova.compute.provider_tree [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.457726] env[65107]: DEBUG oslo_vmware.api [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103288, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.560785} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.458815] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.458815] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.458815] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.458815] env[65107]: INFO nova.compute.manager [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1002.459079] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1002.459079] env[65107]: DEBUG nova.compute.manager [-] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1002.459166] env[65107]: DEBUG nova.network.neutron [-] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1002.459412] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.459948] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.460226] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.507468] env[65107]: DEBUG nova.network.neutron [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updated VIF entry in instance network info cache for port 416a2590-6a42-4a32-a922-a9d5f95786d6. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1002.507914] env[65107]: DEBUG nova.network.neutron [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1002.583264] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.607619] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33d53a18-645a-4378-a7e3-502beb38acbd tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.699s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.671486] env[65107]: DEBUG nova.compute.manager [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1002.671761] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.672681] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f1f839-0a22-4609-a257-8c37dccddc2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.684465] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.684750] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cbebe6e-1c52-4177-ad19-a996cc740641 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.768979] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.768979] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.768979] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Deleting the datastore file [datastore2] 6d5eb77c-c6e8-4ef1-a945-b3485aedd488 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.769336] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f77b2b5-4fb2-46ce-a548-b246988d92ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.785283] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527f8d64-3d12-a2bb-34f2-bfe272d39612, 'name': SearchDatastore_Task, 'duration_secs': 0.020773} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.787402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1002.787922] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.788499] env[65107]: DEBUG oslo_vmware.api [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 1002.788499] env[65107]: value = "task-5103290" [ 1002.788499] env[65107]: _type = "Task" [ 1002.788499] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.789830] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22d7f0a2-f458-43cb-83c5-61fe38b99e2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.803354] env[65107]: DEBUG oslo_vmware.api [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.807313] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1002.807313] env[65107]: value = "task-5103291" [ 1002.807313] env[65107]: _type = "Task" [ 1002.807313] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.814583] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.841474] env[65107]: DEBUG nova.scheduler.client.report [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1003.011175] env[65107]: DEBUG oslo_concurrency.lockutils [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1003.014026] env[65107]: DEBUG nova.compute.manager [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Received event network-vif-deleted-a34ba705-83e3-4da1-ad09-b23e6fc8d54e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1003.014026] env[65107]: DEBUG nova.compute.manager [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Received event network-changed-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1003.014026] env[65107]: DEBUG nova.compute.manager [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Refreshing instance network info cache due to event network-changed-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1003.014026] env[65107]: DEBUG oslo_concurrency.lockutils [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Acquiring lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.014026] env[65107]: DEBUG oslo_concurrency.lockutils [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Acquired lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.014026] env[65107]: DEBUG nova.network.neutron [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Refreshing network info cache for port 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1003.300604] env[65107]: DEBUG nova.network.neutron [-] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1003.315562] env[65107]: DEBUG oslo_vmware.api [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224739} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.315562] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.317110] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.317110] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.317110] env[65107]: INFO nova.compute.manager [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1003.317110] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1003.317110] env[65107]: DEBUG nova.compute.manager [-] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1003.317110] env[65107]: DEBUG nova.network.neutron [-] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1003.317338] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.317820] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.318122] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.335238] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103291, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.351852] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.358486] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.960s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.358486] env[65107]: DEBUG nova.objects.instance [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'resources' on Instance uuid 82e8c093-b718-4d38-9682-ba2e710e4b93 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.394902] env[65107]: INFO nova.scheduler.client.report [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleted allocations for instance 58345821-536a-46ad-af55-92de7552e924 [ 1003.438217] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.518416] env[65107]: WARNING neutronclient.v2_0.client [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.519322] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.519667] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.811017] env[65107]: INFO nova.compute.manager [-] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Took 1.35 seconds to deallocate network for instance. [ 1003.822714] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609672} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.825459] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.825861] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.832465] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.832710] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.835384] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-498170b3-fe8c-4fde-a5eb-c1be2b02d92f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.838589] env[65107]: DEBUG nova.compute.manager [req-0c849acb-9679-48f1-ae52-b435a8c824f8 req-527064b0-1539-4e67-9c39-5ff5f51c67e0 service nova] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Received event network-vif-deleted-355ade04-dfe9-489a-81d1-b7bee4c4833e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1003.846301] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1003.846301] env[65107]: value = "task-5103292" [ 1003.846301] env[65107]: _type = "Task" [ 1003.846301] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.861710] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103292, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.907687] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d24edf86-d068-4a09-9056-0afdce20ac67 tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "58345821-536a-46ad-af55-92de7552e924" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.280s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.911984] env[65107]: WARNING neutronclient.v2_0.client [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.912858] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.913118] env[65107]: WARNING openstack [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.068397] env[65107]: DEBUG nova.network.neutron [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updated VIF entry in instance network info cache for port 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1004.068792] env[65107]: DEBUG nova.network.neutron [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updating instance_info_cache with network_info: [{"id": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "address": "fa:16:3e:4e:cd:29", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfc6b6-0a", "ovs_interfaceid": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.161219] env[65107]: DEBUG nova.network.neutron [-] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.240245] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae7303b-5bbc-49bd-bd6a-f558e4d764d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.249075] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6acc94b-cb9c-440b-a72c-53432f64fb57 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.281059] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c29bd8-cd7b-4e04-8a69-444637f4de6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.289799] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e360f668-810c-45cb-a2f7-f3b4b1a11f05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.295840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.296081] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.296283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.296459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.296622] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.299549] env[65107]: INFO nova.compute.manager [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Terminating instance [ 1004.310208] env[65107]: DEBUG nova.compute.provider_tree [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.324065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.356940] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103292, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093673} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.359143] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.359143] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0153288d-1e0e-40fb-8489-c737749dc13e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.383683] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.384058] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec117570-0cdd-4ef4-85d9-b536b07aa2c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.406660] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1004.406660] env[65107]: value = "task-5103293" [ 1004.406660] env[65107]: _type = "Task" [ 1004.406660] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.416123] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103293, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.578708] env[65107]: DEBUG oslo_concurrency.lockutils [req-50de58a2-3491-4948-accd-1e3132822e63 req-e9ffd613-2869-4369-90e3-c21b7b1be04b service nova] Releasing lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1004.667082] env[65107]: INFO nova.compute.manager [-] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Took 1.35 seconds to deallocate network for instance. [ 1004.814801] env[65107]: DEBUG nova.scheduler.client.report [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1004.819453] env[65107]: DEBUG nova.compute.manager [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1004.819708] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1004.820886] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f0b584-ec54-4fb4-93e8-b03705fbb4c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.829892] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.830158] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-775f0db3-c729-4b53-a360-3956cc4006c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.837700] env[65107]: DEBUG oslo_vmware.api [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 1004.837700] env[65107]: value = "task-5103294" [ 1004.837700] env[65107]: _type = "Task" [ 1004.837700] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.849319] env[65107]: DEBUG oslo_vmware.api [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.918131] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103293, 'name': ReconfigVM_Task, 'duration_secs': 0.326132} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.918561] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.919282] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba2d8fae-4cc2-4424-84a3-34d94bcdb02b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.927286] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1004.927286] env[65107]: value = "task-5103295" [ 1004.927286] env[65107]: _type = "Task" [ 1004.927286] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.937666] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103295, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.082428] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "c5fb4b1c-5540-4645-b996-f397729beb23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.082710] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "c5fb4b1c-5540-4645-b996-f397729beb23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.082899] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "c5fb4b1c-5540-4645-b996-f397729beb23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.083110] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "c5fb4b1c-5540-4645-b996-f397729beb23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.083349] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "c5fb4b1c-5540-4645-b996-f397729beb23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.085973] env[65107]: INFO nova.compute.manager [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Terminating instance [ 1005.175183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.322458] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.969s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.326812] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.589s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.327536] env[65107]: DEBUG nova.objects.instance [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'resources' on Instance uuid d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.354679] env[65107]: DEBUG oslo_vmware.api [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103294, 'name': PowerOffVM_Task, 'duration_secs': 0.258244} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.356214] env[65107]: INFO nova.scheduler.client.report [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted allocations for instance 82e8c093-b718-4d38-9682-ba2e710e4b93 [ 1005.357735] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.358053] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1005.361873] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-04c39eb7-cdae-4be9-9110-9805c5b1252a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.443603] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103295, 'name': Rename_Task, 'duration_secs': 0.163998} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.443916] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.444278] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0b5d742-465c-4233-9a85-32adc9a06442 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.453611] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1005.453611] env[65107]: value = "task-5103297" [ 1005.453611] env[65107]: _type = "Task" [ 1005.453611] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.462728] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.591096] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "refresh_cache-c5fb4b1c-5540-4645-b996-f397729beb23" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.591096] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquired lock "refresh_cache-c5fb4b1c-5540-4645-b996-f397729beb23" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1005.591096] env[65107]: DEBUG nova.network.neutron [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1005.634059] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1005.634389] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1005.634702] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Deleting the datastore file [datastore1] 59ff1aa6-8d0d-4885-90d3-33d623dca8bb {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.635109] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed247bd8-0a6f-42cd-979b-64687980cd50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.644009] env[65107]: DEBUG oslo_vmware.api [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for the task: (returnval){ [ 1005.644009] env[65107]: value = "task-5103298" [ 1005.644009] env[65107]: _type = "Task" [ 1005.644009] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.655265] env[65107]: DEBUG oslo_vmware.api [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.832402] env[65107]: DEBUG nova.objects.instance [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'numa_topology' on Instance uuid d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.862929] env[65107]: DEBUG nova.compute.manager [req-0ff9b2a1-cda6-4ab2-a3b4-4cf68952b7e0 req-ad2f94ad-1905-4fa5-8dd3-cf1c5c89bc7b service nova] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Received event network-vif-deleted-51b0e4fe-384d-4ad6-8341-4525d8b55ffa {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1005.867285] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4c10ccb4-b1a0-4089-9410-4b1f8e894789 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "82e8c093-b718-4d38-9682-ba2e710e4b93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.135s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.967628] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103297, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.094717] env[65107]: WARNING neutronclient.v2_0.client [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.095601] env[65107]: WARNING openstack [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.096251] env[65107]: WARNING openstack [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.159202] env[65107]: DEBUG oslo_vmware.api [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Task: {'id': task-5103298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237325} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.159477] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.159694] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.159957] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.160061] env[65107]: INFO nova.compute.manager [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Took 1.34 seconds to destroy the instance on the hypervisor. [ 1006.160315] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1006.161079] env[65107]: DEBUG nova.compute.manager [-] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1006.161079] env[65107]: DEBUG nova.network.neutron [-] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1006.161079] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.161353] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.161650] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.180940] env[65107]: DEBUG nova.network.neutron [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1006.282526] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.320997] env[65107]: DEBUG nova.network.neutron [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1006.335691] env[65107]: DEBUG nova.objects.base [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1006.465765] env[65107]: DEBUG oslo_vmware.api [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103297, 'name': PowerOnVM_Task, 'duration_secs': 0.569274} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.468773] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.469036] env[65107]: INFO nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Took 8.32 seconds to spawn the instance on the hypervisor. [ 1006.469237] env[65107]: DEBUG nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1006.470513] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7feae26a-ffe0-4579-ad48-1ad8bb004dbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.753233] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb36476-d4a2-47b2-8b02-63cdbdd92293 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.765227] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72715216-5d45-417e-b80f-5f86224f8c18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.798157] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718c3ce7-220c-43ff-98dc-1abbda4fd33e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.807035] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baf8d40-51e2-4209-8c00-1b4b8dd2aadc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.824567] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Releasing lock "refresh_cache-c5fb4b1c-5540-4645-b996-f397729beb23" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.824956] env[65107]: DEBUG nova.compute.manager [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1006.826030] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.826030] env[65107]: DEBUG nova.compute.provider_tree [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.827642] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7fd486-9d9a-47a5-923b-eb314e578ab1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.836463] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.836742] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4a1c35b-c4a3-4a44-b9c8-84abd18219a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.845545] env[65107]: DEBUG oslo_vmware.api [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 1006.845545] env[65107]: value = "task-5103299" [ 1006.845545] env[65107]: _type = "Task" [ 1006.845545] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.856448] env[65107]: DEBUG oslo_vmware.api [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.996161] env[65107]: INFO nova.compute.manager [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Took 39.00 seconds to build instance. [ 1007.045615] env[65107]: DEBUG nova.network.neutron [-] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1007.332880] env[65107]: DEBUG nova.scheduler.client.report [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1007.358030] env[65107]: DEBUG oslo_vmware.api [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103299, 'name': PowerOffVM_Task, 'duration_secs': 0.156854} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.358312] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.358463] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.358658] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a40e9032-70d0-4cf2-999b-4075c91516ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.390425] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.390627] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.390824] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleting the datastore file [datastore2] c5fb4b1c-5540-4645-b996-f397729beb23 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.391644] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f088feb0-6bb7-40a8-952f-b1ffc5da6a00 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.399144] env[65107]: DEBUG oslo_vmware.api [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for the task: (returnval){ [ 1007.399144] env[65107]: value = "task-5103301" [ 1007.399144] env[65107]: _type = "Task" [ 1007.399144] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.410394] env[65107]: DEBUG oslo_vmware.api [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.498246] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e6da290d-bcc8-48ec-9e13-820af5778d46 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.508s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.549235] env[65107]: INFO nova.compute.manager [-] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Took 1.39 seconds to deallocate network for instance. [ 1007.743616] env[65107]: DEBUG nova.compute.manager [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Received event network-changed-416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1007.743726] env[65107]: DEBUG nova.compute.manager [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Refreshing instance network info cache due to event network-changed-416a2590-6a42-4a32-a922-a9d5f95786d6. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1007.743956] env[65107]: DEBUG oslo_concurrency.lockutils [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.744115] env[65107]: DEBUG oslo_concurrency.lockutils [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.744274] env[65107]: DEBUG nova.network.neutron [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Refreshing network info cache for port 416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1007.839747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.513s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.843024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.546s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.843356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.843569] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1007.843945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.109s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.844183] env[65107]: DEBUG nova.objects.instance [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1007.847768] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ad75f6-6bb5-4dc4-b6fe-8f4667bc733c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.860783] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc001f08-e85a-45dc-ba5e-993e5e075ebb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.880637] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.880637] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.883057] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa56d1b5-7982-4ddb-9483-a60e10d0f880 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.902881] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3a32af-f9e5-47b8-af10-f1b26c2a1a28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.914738] env[65107]: DEBUG nova.compute.manager [req-79dd518b-96e0-4f02-aef2-dd7e5662932c req-d3fda9ed-967d-42ef-a44b-220a1a00cd11 service nova] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Received event network-vif-deleted-8f09ac6e-ef41-4d62-b1fd-acf86cfbed59 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1007.922465] env[65107]: DEBUG oslo_vmware.api [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Task: {'id': task-5103301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.30483} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.950448] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1007.950717] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1007.950932] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1007.952029] env[65107]: INFO nova.compute.manager [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1007.952029] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1007.952376] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178343MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1007.955093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.955093] env[65107]: DEBUG nova.compute.manager [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1007.955093] env[65107]: DEBUG nova.network.neutron [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1007.955093] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.955093] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.955093] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.979346] env[65107]: DEBUG nova.network.neutron [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1007.979407] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.057268] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.247140] env[65107]: WARNING neutronclient.v2_0.client [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.247838] env[65107]: WARNING openstack [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.248295] env[65107]: WARNING openstack [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.354140] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b7047553-2c19-4aad-90ee-c218728bb127 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 42.352s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.354960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 17.307s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.355741] env[65107]: INFO nova.compute.manager [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Unshelving [ 1008.384238] env[65107]: WARNING openstack [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.384609] env[65107]: WARNING openstack [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.396070] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1008.454551] env[65107]: WARNING neutronclient.v2_0.client [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.455265] env[65107]: WARNING openstack [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.455614] env[65107]: WARNING openstack [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.483008] env[65107]: DEBUG nova.network.neutron [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.540399] env[65107]: DEBUG nova.network.neutron [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updated VIF entry in instance network info cache for port 416a2590-6a42-4a32-a922-a9d5f95786d6. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1008.540792] env[65107]: DEBUG nova.network.neutron [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.855622] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5656476d-3e53-43be-867f-ab12722bbd1f tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.856948] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.518s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.858467] env[65107]: INFO nova.compute.claims [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.919186] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.985851] env[65107]: INFO nova.compute.manager [-] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Took 1.03 seconds to deallocate network for instance. [ 1009.043770] env[65107]: DEBUG oslo_concurrency.lockutils [req-7caed832-a0c6-4cfb-9acb-d66478d1ea4a req-6c09d676-a4e2-448e-8f6a-19fe22a1cf90 service nova] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1009.370127] env[65107]: DEBUG nova.compute.utils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1009.493713] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.874184] env[65107]: INFO nova.virt.block_device [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Booting with volume b9b6229b-f202-42ba-8d20-88f1554891f7 at /dev/sdb [ 1009.917450] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c698d2d-cc6d-41d1-91c8-0d6689bf7205 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.932255] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b926740-fa6f-40fb-9395-abe086df8e34 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.973871] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60920bcc-63ca-4f64-a6f7-c709b8b89349 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.984068] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ef0b5f-a2c5-48d4-921a-c5e7508aaae0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.026156] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0888ec4a-918d-4a19-b130-30fed08dfb92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.034109] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77680d32-ca02-4051-81d3-9426c8cfaa19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.054171] env[65107]: DEBUG nova.virt.block_device [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating existing volume attachment record: c9b36cfe-71bf-431b-94dd-0dccd26fcfcf {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1010.219055] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42474f75-0cbf-49a9-b476-fe5c0ecd54e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.226634] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5be1a4c-3a83-4fe6-bcf7-683d7adc204c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.260287] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ac4368-8819-4296-9167-9431a3d9ac56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.268016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277e8058-9feb-49a8-a87f-b910e1a25626 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.282342] env[65107]: DEBUG nova.compute.provider_tree [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.305068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.305308] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.785367] env[65107]: DEBUG nova.scheduler.client.report [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.808342] env[65107]: DEBUG nova.compute.utils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1011.290476] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.290981] env[65107]: DEBUG nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1011.293621] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.718s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.295336] env[65107]: INFO nova.compute.claims [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.310581] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.799817] env[65107]: DEBUG nova.compute.utils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1011.804852] env[65107]: DEBUG nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Not allocating networking since 'none' was specified. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1012.305614] env[65107]: DEBUG nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1012.376161] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.376525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.376824] env[65107]: INFO nova.compute.manager [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Attaching volume e8465018-2a5c-45af-b181-4202b8bd8fce to /dev/sdb [ 1012.414584] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df650fd1-2486-4e17-b926-4429d1375174 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.422148] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589ce771-9deb-437a-a77f-39b45eff68a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.440660] env[65107]: DEBUG nova.virt.block_device [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating existing volume attachment record: 6e621168-6821-4b17-8a57-b969beca6f93 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1012.609640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d90100-fb42-4229-9a80-2d4f8fed9b78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.617198] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1508fa-48c8-4b50-964f-3a0a0d8b2a66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.647252] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2ebd35-5082-4d33-80c0-74bdf2180cd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.655185] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9ce405-c335-45c7-8eb7-35650726fde6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.669164] env[65107]: DEBUG nova.compute.provider_tree [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.175907] env[65107]: DEBUG nova.scheduler.client.report [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1013.318871] env[65107]: DEBUG nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1013.347201] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1013.347549] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1013.347733] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1013.347916] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1013.348077] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1013.348224] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1013.348463] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.348652] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1013.348843] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1013.349049] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1013.349235] env[65107]: DEBUG nova.virt.hardware [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1013.350247] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96b04f3-595b-4d46-a665-4fc32bd234ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.359105] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c401ce65-170a-41e7-bcda-6de4fce2bb21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.374224] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.379754] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Creating folder: Project (8ff69cbe020d4f28a1374e9ce3721416). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1013.380077] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ead6387-a34e-4bac-8726-6652650bc3e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.392038] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Created folder: Project (8ff69cbe020d4f28a1374e9ce3721416) in parent group-v992574. [ 1013.392255] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Creating folder: Instances. Parent ref: group-v992832. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1013.392515] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-133abdc7-e02b-4c06-9f32-2a0d1c9e128f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.401324] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Created folder: Instances in parent group-v992832. [ 1013.401557] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1013.401750] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.401950] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-609cc94b-d007-402c-9855-0b70acaaf319 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.418554] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.418554] env[65107]: value = "task-5103311" [ 1013.418554] env[65107]: _type = "Task" [ 1013.418554] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.426183] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103311, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.683243] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.683795] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1013.687149] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.964s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.687149] env[65107]: DEBUG nova.objects.instance [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lazy-loading 'resources' on Instance uuid 13c96e2f-46de-46e0-a505-60dfec9e95ca {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.929434] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103311, 'name': CreateVM_Task, 'duration_secs': 0.350183} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.929641] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.930118] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.930287] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.930624] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1013.931289] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee25dbc5-a081-4436-a4c7-897ced5c23fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.937384] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1013.937384] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d439a-7c3c-4c4f-2033-ddb9d08085a5" [ 1013.937384] env[65107]: _type = "Task" [ 1013.937384] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.947617] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d439a-7c3c-4c4f-2033-ddb9d08085a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.195399] env[65107]: DEBUG nova.compute.utils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1014.198092] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1014.199045] env[65107]: DEBUG nova.network.neutron [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1014.199045] env[65107]: WARNING neutronclient.v2_0.client [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.199391] env[65107]: WARNING neutronclient.v2_0.client [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.200202] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.202806] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.268526] env[65107]: DEBUG nova.policy [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ccab6477dd44878f78a9d6428d3cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4778fe9152224fd29d1f6220a19b5a36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1014.453135] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d439a-7c3c-4c4f-2033-ddb9d08085a5, 'name': SearchDatastore_Task, 'duration_secs': 0.012082} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.453135] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.453135] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.453477] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.453620] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.453801] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.454086] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03a8508a-32ed-4dc0-9d6c-1cce99711263 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.468378] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.468378] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.471511] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c83c483-b794-4392-aab7-fc89ccd7f59e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.478783] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1014.478783] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528064ff-4cd3-3b6b-4f62-4404d13839de" [ 1014.478783] env[65107]: _type = "Task" [ 1014.478783] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.486342] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528064ff-4cd3-3b6b-4f62-4404d13839de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.539232] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff674669-154e-4fe8-9ca1-6f380e67ced9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.549046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c07dd8d-07c8-44d0-a81b-2e4460ad6f8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.580845] env[65107]: DEBUG nova.network.neutron [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Successfully created port: d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1014.583533] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb170ce-8449-4adf-b4f8-c6aeb75a3099 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.591283] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbf97e2-3ede-4ebf-a45f-a9dc1bc6471a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.605011] env[65107]: DEBUG nova.compute.provider_tree [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.713223] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1014.988834] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528064ff-4cd3-3b6b-4f62-4404d13839de, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.989668] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71c4c249-a845-4f94-ad7d-48329c7a3bad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.995046] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1014.995046] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52cf53b1-2633-4ea0-8e23-c8fd419c67f5" [ 1014.995046] env[65107]: _type = "Task" [ 1014.995046] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.003380] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52cf53b1-2633-4ea0-8e23-c8fd419c67f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.108126] env[65107]: DEBUG nova.scheduler.client.report [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1015.507263] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52cf53b1-2633-4ea0-8e23-c8fd419c67f5, 'name': SearchDatastore_Task, 'duration_secs': 0.009967} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.507571] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.507792] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.508087] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff433971-f909-4b85-bbd5-f729a7d08b82 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.516083] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1015.516083] env[65107]: value = "task-5103313" [ 1015.516083] env[65107]: _type = "Task" [ 1015.516083] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.527812] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.618455] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.623330] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.389s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.623557] env[65107]: DEBUG nova.objects.instance [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1015.649228] env[65107]: INFO nova.scheduler.client.report [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Deleted allocations for instance 13c96e2f-46de-46e0-a505-60dfec9e95ca [ 1015.722440] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1015.751979] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1015.752313] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1015.752504] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1015.752727] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1015.752904] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1015.753087] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1015.753304] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.753465] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1015.753630] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1015.753883] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1015.754076] env[65107]: DEBUG nova.virt.hardware [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1015.755023] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2505590-4bb6-4bd6-be15-949fde875c72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.766175] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70ee5e3-f548-4622-a902-4d0bf08f870e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.019273] env[65107]: DEBUG nova.compute.manager [req-1933c0e9-a3d0-498a-ac54-aa3dd50a5857 req-3a7b59b5-ebb1-47db-a504-af8d9c36d94f service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Received event network-vif-plugged-d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1016.019273] env[65107]: DEBUG oslo_concurrency.lockutils [req-1933c0e9-a3d0-498a-ac54-aa3dd50a5857 req-3a7b59b5-ebb1-47db-a504-af8d9c36d94f service nova] Acquiring lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.019688] env[65107]: DEBUG oslo_concurrency.lockutils [req-1933c0e9-a3d0-498a-ac54-aa3dd50a5857 req-3a7b59b5-ebb1-47db-a504-af8d9c36d94f service nova] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.019959] env[65107]: DEBUG oslo_concurrency.lockutils [req-1933c0e9-a3d0-498a-ac54-aa3dd50a5857 req-3a7b59b5-ebb1-47db-a504-af8d9c36d94f service nova] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.020260] env[65107]: DEBUG nova.compute.manager [req-1933c0e9-a3d0-498a-ac54-aa3dd50a5857 req-3a7b59b5-ebb1-47db-a504-af8d9c36d94f service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] No waiting events found dispatching network-vif-plugged-d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1016.020622] env[65107]: WARNING nova.compute.manager [req-1933c0e9-a3d0-498a-ac54-aa3dd50a5857 req-3a7b59b5-ebb1-47db-a504-af8d9c36d94f service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Received unexpected event network-vif-plugged-d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 for instance with vm_state building and task_state spawning. [ 1016.032736] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48974} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.033038] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.033206] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.033456] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db63ff89-bb99-4847-9865-f47b4efa60c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.040212] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1016.040212] env[65107]: value = "task-5103314" [ 1016.040212] env[65107]: _type = "Task" [ 1016.040212] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.050788] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103314, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.116940] env[65107]: DEBUG nova.network.neutron [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Successfully updated port: d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1016.156904] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01b43b5c-8726-4376-b75b-afcdd6caaece tempest-ServersListShow298Test-817192361 tempest-ServersListShow298Test-817192361-project-member] Lock "13c96e2f-46de-46e0-a505-60dfec9e95ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.740s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.177497] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.555559] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103314, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070017} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.556549] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.558297] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d9e32a-cd0c-4ec4-9c9d-228da0e84f47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.593390] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.593860] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdef003b-5cc9-4d78-9ee4-474dfe410b65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.622135] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-b4f7d003-59fb-4ef4-b5f6-234d5154d198" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.622135] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-b4f7d003-59fb-4ef4-b5f6-234d5154d198" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.622333] env[65107]: DEBUG nova.network.neutron [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1016.630435] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1016.630435] env[65107]: value = "task-5103315" [ 1016.630435] env[65107]: _type = "Task" [ 1016.630435] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.636604] env[65107]: DEBUG oslo_concurrency.lockutils [None req-386b206a-6295-43d1-bd2d-766a93414267 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.637685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.365s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.640046] env[65107]: INFO nova.compute.claims [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.649156] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.985983] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1016.986420] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992831', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'name': 'volume-e8465018-2a5c-45af-b181-4202b8bd8fce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e12cb11f-7d6f-47d6-a31e-bf2654174d38', 'attached_at': '', 'detached_at': '', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'serial': 'e8465018-2a5c-45af-b181-4202b8bd8fce'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1016.987657] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea18a1e-daf8-4c02-9995-f34aecb92de1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.011026] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a0082e-74b2-488a-878e-8042afa57d7d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.038850] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] volume-e8465018-2a5c-45af-b181-4202b8bd8fce/volume-e8465018-2a5c-45af-b181-4202b8bd8fce.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.039203] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f9ea2cc-0307-4636-ad09-1463471377fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.064282] env[65107]: DEBUG oslo_vmware.api [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1017.064282] env[65107]: value = "task-5103316" [ 1017.064282] env[65107]: _type = "Task" [ 1017.064282] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.075200] env[65107]: DEBUG oslo_vmware.api [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103316, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.125551] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.126089] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.148557] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.183692] env[65107]: DEBUG nova.network.neutron [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1017.205470] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.206079] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.280284] env[65107]: WARNING neutronclient.v2_0.client [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.280965] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.281512] env[65107]: WARNING openstack [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.378752] env[65107]: DEBUG nova.network.neutron [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Updating instance_info_cache with network_info: [{"id": "d252ac0c-f7eb-4a04-8af4-7b1eb031aa93", "address": "fa:16:3e:c6:bb:97", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd252ac0c-f7", "ovs_interfaceid": "d252ac0c-f7eb-4a04-8af4-7b1eb031aa93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.575025] env[65107]: DEBUG oslo_vmware.api [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103316, 'name': ReconfigVM_Task, 'duration_secs': 0.386342} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.575531] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfigured VM instance instance-00000050 to attach disk [datastore1] volume-e8465018-2a5c-45af-b181-4202b8bd8fce/volume-e8465018-2a5c-45af-b181-4202b8bd8fce.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.580346] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13c58a06-7475-45bb-baa5-b20c60cda3a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.595550] env[65107]: DEBUG oslo_vmware.api [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1017.595550] env[65107]: value = "task-5103317" [ 1017.595550] env[65107]: _type = "Task" [ 1017.595550] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.605009] env[65107]: DEBUG oslo_vmware.api [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.644949] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103315, 'name': ReconfigVM_Task, 'duration_secs': 0.617992} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.644949] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Reconfigured VM instance instance-0000005a to attach disk [datastore1] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.646340] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-522d9cc9-43b1-478a-bc89-d668f6db7aa1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.655671] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1017.655671] env[65107]: value = "task-5103318" [ 1017.655671] env[65107]: _type = "Task" [ 1017.655671] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.665637] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103318, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.882020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-b4f7d003-59fb-4ef4-b5f6-234d5154d198" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.882562] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Instance network_info: |[{"id": "d252ac0c-f7eb-4a04-8af4-7b1eb031aa93", "address": "fa:16:3e:c6:bb:97", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd252ac0c-f7", "ovs_interfaceid": "d252ac0c-f7eb-4a04-8af4-7b1eb031aa93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1017.883066] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:bb:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd252ac0c-f7eb-4a04-8af4-7b1eb031aa93', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.890970] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1017.894201] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.894688] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97d91dbc-eb51-4e2e-bc54-b9b8312d7438 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.918398] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.918398] env[65107]: value = "task-5103319" [ 1017.918398] env[65107]: _type = "Task" [ 1017.918398] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.927397] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103319, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.947100] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b4bf43-9ff1-4b4c-abb5-a9509be5dd48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.955571] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3b71e0-8a1d-42e2-910c-169ab74c82eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.989312] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250f85e5-caf0-4a50-9dc3-430bae1bb49c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.997941] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610b36bd-3e7d-4326-9916-59788787a907 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.012280] env[65107]: DEBUG nova.compute.provider_tree [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.051169] env[65107]: DEBUG nova.compute.manager [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Received event network-changed-d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1018.051169] env[65107]: DEBUG nova.compute.manager [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Refreshing instance network info cache due to event network-changed-d252ac0c-f7eb-4a04-8af4-7b1eb031aa93. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1018.051684] env[65107]: DEBUG oslo_concurrency.lockutils [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Acquiring lock "refresh_cache-b4f7d003-59fb-4ef4-b5f6-234d5154d198" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.051684] env[65107]: DEBUG oslo_concurrency.lockutils [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Acquired lock "refresh_cache-b4f7d003-59fb-4ef4-b5f6-234d5154d198" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.051841] env[65107]: DEBUG nova.network.neutron [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Refreshing network info cache for port d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1018.108654] env[65107]: DEBUG oslo_vmware.api [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103317, 'name': ReconfigVM_Task, 'duration_secs': 0.150013} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.108654] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992831', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'name': 'volume-e8465018-2a5c-45af-b181-4202b8bd8fce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e12cb11f-7d6f-47d6-a31e-bf2654174d38', 'attached_at': '', 'detached_at': '', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'serial': 'e8465018-2a5c-45af-b181-4202b8bd8fce'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1018.167018] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103318, 'name': Rename_Task, 'duration_secs': 0.149147} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.167436] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1018.167739] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6d1b0f3-b0d0-47fc-b9aa-11a3f3bb615a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.175515] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1018.175515] env[65107]: value = "task-5103320" [ 1018.175515] env[65107]: _type = "Task" [ 1018.175515] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.185196] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103320, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.428803] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103319, 'name': CreateVM_Task, 'duration_secs': 0.34326} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.428987] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.429612] env[65107]: WARNING neutronclient.v2_0.client [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.429900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.430162] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.430414] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1018.431017] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad53afe8-4423-441f-930b-2b1306dafdc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.436281] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1018.436281] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a4d359-1d2d-c91e-2dc8-f57176138f2c" [ 1018.436281] env[65107]: _type = "Task" [ 1018.436281] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.444743] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a4d359-1d2d-c91e-2dc8-f57176138f2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.516092] env[65107]: DEBUG nova.scheduler.client.report [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1018.555947] env[65107]: WARNING neutronclient.v2_0.client [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.556414] env[65107]: WARNING openstack [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.556759] env[65107]: WARNING openstack [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.681880] env[65107]: WARNING openstack [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.682315] env[65107]: WARNING openstack [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.695314] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103320, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.743586] env[65107]: WARNING neutronclient.v2_0.client [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.744509] env[65107]: WARNING openstack [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.744652] env[65107]: WARNING openstack [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.831833] env[65107]: DEBUG nova.network.neutron [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Updated VIF entry in instance network info cache for port d252ac0c-f7eb-4a04-8af4-7b1eb031aa93. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1018.832215] env[65107]: DEBUG nova.network.neutron [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Updating instance_info_cache with network_info: [{"id": "d252ac0c-f7eb-4a04-8af4-7b1eb031aa93", "address": "fa:16:3e:c6:bb:97", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd252ac0c-f7", "ovs_interfaceid": "d252ac0c-f7eb-4a04-8af4-7b1eb031aa93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1018.946947] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a4d359-1d2d-c91e-2dc8-f57176138f2c, 'name': SearchDatastore_Task, 'duration_secs': 0.016349} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.947295] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.947555] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.947793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.947941] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.948137] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.948474] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7bc31a8-3ea8-499c-bd0d-56962fb6486e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.957098] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.957302] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.957980] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6803b31a-3df7-4343-ad67-45edd0d2b305 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.963375] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1018.963375] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522440e8-b2f2-c97a-e7f4-ad86f2aa45e4" [ 1018.963375] env[65107]: _type = "Task" [ 1018.963375] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.971584] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522440e8-b2f2-c97a-e7f4-ad86f2aa45e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.024365] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.024922] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1019.027677] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.449s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.027841] env[65107]: DEBUG nova.objects.instance [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'resources' on Instance uuid 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.152222] env[65107]: DEBUG nova.objects.instance [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'flavor' on Instance uuid e12cb11f-7d6f-47d6-a31e-bf2654174d38 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.186873] env[65107]: DEBUG oslo_vmware.api [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103320, 'name': PowerOnVM_Task, 'duration_secs': 0.680113} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.187173] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1019.187370] env[65107]: INFO nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Took 5.87 seconds to spawn the instance on the hypervisor. [ 1019.187548] env[65107]: DEBUG nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1019.188341] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9b9216-9e07-4df3-bac5-9e3af13c99a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.335796] env[65107]: DEBUG oslo_concurrency.lockutils [req-b1c96432-5a68-46b0-bf4d-294e5861e948 req-41177fb8-91ec-4d95-a2c2-2f10d3c304e7 service nova] Releasing lock "refresh_cache-b4f7d003-59fb-4ef4-b5f6-234d5154d198" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.475790] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522440e8-b2f2-c97a-e7f4-ad86f2aa45e4, 'name': SearchDatastore_Task, 'duration_secs': 0.009197} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.476975] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a4bf69-f604-4721-b647-84c2204aa58d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.483777] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1019.483777] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5242dd10-9a8f-71f1-d5f6-4523dcddcc3d" [ 1019.483777] env[65107]: _type = "Task" [ 1019.483777] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.492693] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5242dd10-9a8f-71f1-d5f6-4523dcddcc3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.534035] env[65107]: DEBUG nova.compute.utils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1019.535748] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1019.535871] env[65107]: DEBUG nova.network.neutron [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1019.536182] env[65107]: WARNING neutronclient.v2_0.client [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.536545] env[65107]: WARNING neutronclient.v2_0.client [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.537104] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.537447] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.588019] env[65107]: DEBUG nova.policy [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '365eed31877241e58fc860ed5620d5f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c805fe68d0fd4147b9b4b7fdc9d6a6d0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1019.658440] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a7c8ff8-3940-40e3-9508-369d5c7e0f4e tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.282s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.710554] env[65107]: INFO nova.compute.manager [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Took 26.40 seconds to build instance. [ 1019.855598] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e742de15-1288-4d26-a7ba-ef7032fec275 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.866427] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9860e99c-61f1-40d5-8acb-4f66f49b4cb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.897454] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e2c9c5-3ae1-4439-8d5d-5ee136c6e8ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.905999] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3426cf75-d851-4d8e-9803-b8e2656b06a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.912181] env[65107]: DEBUG nova.network.neutron [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Successfully created port: d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1019.923764] env[65107]: INFO nova.compute.manager [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Rescuing [ 1019.924344] env[65107]: DEBUG oslo_concurrency.lockutils [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.924344] env[65107]: DEBUG oslo_concurrency.lockutils [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.924344] env[65107]: DEBUG nova.network.neutron [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1019.927645] env[65107]: DEBUG nova.compute.provider_tree [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.995181] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5242dd10-9a8f-71f1-d5f6-4523dcddcc3d, 'name': SearchDatastore_Task, 'duration_secs': 0.009797} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.995798] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.995798] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] b4f7d003-59fb-4ef4-b5f6-234d5154d198/b4f7d003-59fb-4ef4-b5f6-234d5154d198.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.995989] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f682895e-767e-4baf-84a7-1c2eb498d966 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.003676] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1020.003676] env[65107]: value = "task-5103321" [ 1020.003676] env[65107]: _type = "Task" [ 1020.003676] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.013399] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.046479] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1020.213384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a612328c-8d28-45e0-ab00-97fd6cfe4476 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "a3162257-ac6b-4468-a573-3611bba69c46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.911s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.429421] env[65107]: WARNING neutronclient.v2_0.client [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.431099] env[65107]: WARNING openstack [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.431941] env[65107]: WARNING openstack [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.448117] env[65107]: DEBUG nova.scheduler.client.report [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1020.515144] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103321, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.841484] env[65107]: INFO nova.compute.manager [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Rebuilding instance [ 1020.857299] env[65107]: WARNING openstack [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.857782] env[65107]: WARNING openstack [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.904112] env[65107]: DEBUG nova.compute.manager [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1020.905155] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e909db7f-99e8-4eef-97fa-27995a2991a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.958021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.929s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.959870] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.813s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.961814] env[65107]: INFO nova.compute.claims [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.977729] env[65107]: INFO nova.scheduler.client.report [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocations for instance 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21 [ 1020.982309] env[65107]: WARNING neutronclient.v2_0.client [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.982829] env[65107]: WARNING openstack [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.983796] env[65107]: WARNING openstack [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.020170] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534866} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.020170] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] b4f7d003-59fb-4ef4-b5f6-234d5154d198/b4f7d003-59fb-4ef4-b5f6-234d5154d198.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1021.020170] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.020170] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2bd2992-6cec-41f6-8c05-963f9c0f4ab7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.028511] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1021.028511] env[65107]: value = "task-5103322" [ 1021.028511] env[65107]: _type = "Task" [ 1021.028511] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.038944] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103322, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.057456] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1021.091436] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1021.091617] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1021.091877] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1021.092256] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1021.092256] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1021.092435] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1021.092646] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.092833] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1021.093028] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1021.093200] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1021.093422] env[65107]: DEBUG nova.virt.hardware [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1021.094632] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec312be5-9433-4f92-8c40-5a23b8d14296 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.105492] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91eaf35-5955-46f2-97ca-5944f1ecb7ec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.128167] env[65107]: DEBUG nova.network.neutron [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.393574] env[65107]: DEBUG nova.compute.manager [req-5e99b908-ea36-4ad5-b646-049aecad6d08 req-bf83e5ea-66c8-45b9-a421-cf26c493a3cb service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Received event network-vif-plugged-d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1021.393940] env[65107]: DEBUG oslo_concurrency.lockutils [req-5e99b908-ea36-4ad5-b646-049aecad6d08 req-bf83e5ea-66c8-45b9-a421-cf26c493a3cb service nova] Acquiring lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1021.394236] env[65107]: DEBUG oslo_concurrency.lockutils [req-5e99b908-ea36-4ad5-b646-049aecad6d08 req-bf83e5ea-66c8-45b9-a421-cf26c493a3cb service nova] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.394482] env[65107]: DEBUG oslo_concurrency.lockutils [req-5e99b908-ea36-4ad5-b646-049aecad6d08 req-bf83e5ea-66c8-45b9-a421-cf26c493a3cb service nova] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.394711] env[65107]: DEBUG nova.compute.manager [req-5e99b908-ea36-4ad5-b646-049aecad6d08 req-bf83e5ea-66c8-45b9-a421-cf26c493a3cb service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] No waiting events found dispatching network-vif-plugged-d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1021.395214] env[65107]: WARNING nova.compute.manager [req-5e99b908-ea36-4ad5-b646-049aecad6d08 req-bf83e5ea-66c8-45b9-a421-cf26c493a3cb service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Received unexpected event network-vif-plugged-d88e66c9-ba7c-452d-8e44-8c9603f2b258 for instance with vm_state building and task_state spawning. [ 1021.482023] env[65107]: DEBUG nova.network.neutron [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Successfully updated port: d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1021.494992] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7da009b5-f093-4f42-b344-dacc1167af45 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.215s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.539335] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103322, 'name': ExtendVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.631143] env[65107]: DEBUG oslo_concurrency.lockutils [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.927462] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.927818] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-446f6c6f-3e13-40aa-a035-3d15be76ac66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.936118] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1021.936118] env[65107]: value = "task-5103323" [ 1021.936118] env[65107]: _type = "Task" [ 1021.936118] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.946465] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.984528] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.984824] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.985061] env[65107]: DEBUG nova.network.neutron [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1022.042884] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103322, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.514451} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.043190] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.044016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bd638a-6e6f-4e54-8c48-ded7bea9dd30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.071367] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] b4f7d003-59fb-4ef4-b5f6-234d5154d198/b4f7d003-59fb-4ef4-b5f6-234d5154d198.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.074519] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e17dda5-5d65-44f0-b8a1-b7e9f2937bfa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.096744] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1022.096744] env[65107]: value = "task-5103324" [ 1022.096744] env[65107]: _type = "Task" [ 1022.096744] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.108958] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103324, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.292808] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c477f5f6-295b-4c92-814a-10ef749f11b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.300988] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bd2d87-f8d3-4d44-adad-af61ae2ad1ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.335859] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4780d68f-dd30-4094-90e6-6baf2ccd534a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.344863] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeecbc53-a43b-4828-a2d3-8c98617d83a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.360257] env[65107]: DEBUG nova.compute.provider_tree [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.446074] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103323, 'name': PowerOffVM_Task, 'duration_secs': 0.310741} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.446074] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.446571] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.447484] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7159456b-666a-4cae-a4e5-fb3adb0f89cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.454067] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.454311] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca2430ea-fbf6-4054-9d5f-9154cb2db7aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.481748] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.481960] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.482158] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Deleting the datastore file [datastore1] a3162257-ac6b-4468-a573-3611bba69c46 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.482570] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85af775d-5ef6-4d56-99de-477e76f86e37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.487881] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.488355] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.498079] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1022.498079] env[65107]: value = "task-5103326" [ 1022.498079] env[65107]: _type = "Task" [ 1022.498079] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.507050] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.526350] env[65107]: DEBUG nova.network.neutron [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1022.546467] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.546847] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.606947] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103324, 'name': ReconfigVM_Task, 'duration_secs': 0.322025} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.607272] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Reconfigured VM instance instance-0000005b to attach disk [datastore1] b4f7d003-59fb-4ef4-b5f6-234d5154d198/b4f7d003-59fb-4ef4-b5f6-234d5154d198.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.607932] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17512467-86a2-476b-873a-8a8f8fab78cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.612460] env[65107]: WARNING neutronclient.v2_0.client [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1022.613135] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.613500] env[65107]: WARNING openstack [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.622049] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1022.622049] env[65107]: value = "task-5103327" [ 1022.622049] env[65107]: _type = "Task" [ 1022.622049] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.630572] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103327, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.706014] env[65107]: DEBUG nova.network.neutron [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [{"id": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "address": "fa:16:3e:c6:73:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd88e66c9-ba", "ovs_interfaceid": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1022.864506] env[65107]: DEBUG nova.scheduler.client.report [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1023.008698] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103326, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087432} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.009029] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.009089] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.009326] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.135750] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103327, 'name': Rename_Task, 'duration_secs': 0.139462} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.135750] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.135750] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c22e1b33-671d-47e0-a299-0e6f023f6d96 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.142065] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1023.142065] env[65107]: value = "task-5103328" [ 1023.142065] env[65107]: _type = "Task" [ 1023.142065] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.152188] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103328, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.171231] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.171442] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e98b0f5-266b-4caa-9d88-60e8a4289a2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.180152] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1023.180152] env[65107]: value = "task-5103329" [ 1023.180152] env[65107]: _type = "Task" [ 1023.180152] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.189075] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.208755] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.209167] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Instance network_info: |[{"id": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "address": "fa:16:3e:c6:73:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd88e66c9-ba", "ovs_interfaceid": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1023.209637] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:73:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd88e66c9-ba7c-452d-8e44-8c9603f2b258', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.217452] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1023.217606] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1023.217754] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd243711-d2a0-44a4-af44-0c833e6357a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.238525] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.238525] env[65107]: value = "task-5103330" [ 1023.238525] env[65107]: _type = "Task" [ 1023.238525] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.247289] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103330, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.370115] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.371031] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1023.373821] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.063s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.375636] env[65107]: INFO nova.compute.claims [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.485490] env[65107]: DEBUG nova.compute.manager [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Received event network-changed-d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1023.485719] env[65107]: DEBUG nova.compute.manager [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Refreshing instance network info cache due to event network-changed-d88e66c9-ba7c-452d-8e44-8c9603f2b258. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1023.485938] env[65107]: DEBUG oslo_concurrency.lockutils [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Acquiring lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.486135] env[65107]: DEBUG oslo_concurrency.lockutils [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Acquired lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.486403] env[65107]: DEBUG nova.network.neutron [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Refreshing network info cache for port d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1023.653081] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103328, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.692402] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103329, 'name': PowerOffVM_Task, 'duration_secs': 0.261147} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.692402] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.692945] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a0b571-0729-4e84-98cc-0cb51cf97267 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.715314] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ffaf3c-957a-4f66-b38d-8c88b0deb8aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.750872] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103330, 'name': CreateVM_Task, 'duration_secs': 0.344342} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.751079] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1023.751582] env[65107]: WARNING neutronclient.v2_0.client [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.752707] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.752707] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.752707] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1023.752973] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc698bdc-d62f-4a83-b8a1-aa55fb703789 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.756334] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.756572] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b79010c8-1f81-41f3-a23a-36e18fb9e095 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.762763] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1023.762763] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fa76db-7f2d-3124-517b-d97f8a64b97d" [ 1023.762763] env[65107]: _type = "Task" [ 1023.762763] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.769433] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1023.769433] env[65107]: value = "task-5103331" [ 1023.769433] env[65107]: _type = "Task" [ 1023.769433] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.776181] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fa76db-7f2d-3124-517b-d97f8a64b97d, 'name': SearchDatastore_Task, 'duration_secs': 0.011786} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.776793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.777054] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.777283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.777428] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.777602] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.777851] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59d68828-227e-48d6-ada3-24b10b344d1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.783466] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1023.783652] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.783871] env[65107]: DEBUG oslo_concurrency.lockutils [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.789169] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.789345] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.790030] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e05d2a7b-c289-4b41-bdd5-6f88376233c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.795071] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1023.795071] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d7e31-f7c3-f164-5411-af9c2ecf684b" [ 1023.795071] env[65107]: _type = "Task" [ 1023.795071] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.802733] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d7e31-f7c3-f164-5411-af9c2ecf684b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.880916] env[65107]: DEBUG nova.compute.utils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1023.884423] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1023.884624] env[65107]: DEBUG nova.network.neutron [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1023.884943] env[65107]: WARNING neutronclient.v2_0.client [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.885252] env[65107]: WARNING neutronclient.v2_0.client [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.885889] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.886428] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.948855] env[65107]: DEBUG nova.policy [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32155078a41249a297470fd4083860df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f9188f3e35e4d06a977e2180918c616', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1023.990935] env[65107]: WARNING neutronclient.v2_0.client [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.991612] env[65107]: WARNING openstack [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.991955] env[65107]: WARNING openstack [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.047624] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1024.047624] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1024.048458] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1024.048458] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1024.048458] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1024.048458] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1024.048621] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.048760] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1024.048934] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1024.049283] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1024.049492] env[65107]: DEBUG nova.virt.hardware [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1024.050986] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389d1a0b-3972-4ca7-86c5-25b29a3a5bc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.062212] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a88f718-d69c-4943-9ec8-da2d8d5a16fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.080373] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance VIF info [] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.086109] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1024.088926] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1024.089452] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-369af490-66d2-49eb-9099-7293a4ba488b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.110505] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.110505] env[65107]: value = "task-5103332" [ 1024.110505] env[65107]: _type = "Task" [ 1024.110505] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.122695] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103332, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.146125] env[65107]: WARNING openstack [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.146563] env[65107]: WARNING openstack [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.167376] env[65107]: DEBUG oslo_vmware.api [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103328, 'name': PowerOnVM_Task, 'duration_secs': 0.578709} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.168135] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.168135] env[65107]: INFO nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1024.168135] env[65107]: DEBUG nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1024.169875] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387205fe-2424-4490-8623-0b2f1f2071f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.253936] env[65107]: WARNING neutronclient.v2_0.client [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.254638] env[65107]: WARNING openstack [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.254999] env[65107]: WARNING openstack [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.266827] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2632b756-8614-40f5-9558-a2d7723129ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.275814] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864e2d36-23c2-43bf-a0ed-7d3308609d28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.312496] env[65107]: DEBUG nova.network.neutron [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Successfully created port: 4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1024.318483] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd64b300-8679-43cb-8d3a-01e58f898b63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.329142] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524d7e31-f7c3-f164-5411-af9c2ecf684b, 'name': SearchDatastore_Task, 'duration_secs': 0.009107} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.332825] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95a33296-202d-41a7-96be-3103ca01bd8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.336767] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2558c26-cd34-435c-81e2-3db0bf5a4525 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.350050] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1024.350050] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524f85c7-3287-027a-91c4-bea6140a18ea" [ 1024.350050] env[65107]: _type = "Task" [ 1024.350050] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.358780] env[65107]: DEBUG nova.compute.provider_tree [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.369115] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524f85c7-3287-027a-91c4-bea6140a18ea, 'name': SearchDatastore_Task, 'duration_secs': 0.010354} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.370236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.370930] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65/00574b77-dad6-4f0a-bbcc-20a2b4a5df65.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.370930] env[65107]: DEBUG oslo_concurrency.lockutils [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1024.370930] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.371268] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ad691d3-c8c3-4ebc-801e-3f5043898cf4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.373538] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5299dd2b-8796-4bcb-9da7-a6b05efc2004 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.381358] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1024.381358] env[65107]: value = "task-5103333" [ 1024.381358] env[65107]: _type = "Task" [ 1024.381358] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.387468] env[65107]: DEBUG nova.network.neutron [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updated VIF entry in instance network info cache for port d88e66c9-ba7c-452d-8e44-8c9603f2b258. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1024.387836] env[65107]: DEBUG nova.network.neutron [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [{"id": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "address": "fa:16:3e:c6:73:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd88e66c9-ba", "ovs_interfaceid": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1024.390440] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1024.393248] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.393423] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1024.394822] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3e3d394-0fdc-4e50-bdb7-a5c9881ce70c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.401358] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103333, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.405821] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1024.405821] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5253f17c-1602-7b42-e64b-c2c415423f30" [ 1024.405821] env[65107]: _type = "Task" [ 1024.405821] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.419241] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5253f17c-1602-7b42-e64b-c2c415423f30, 'name': SearchDatastore_Task, 'duration_secs': 0.009847} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.420116] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-933d9dd0-14bf-499f-9032-9e6e765d65f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.425991] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1024.425991] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed889b-5ceb-fe52-baee-9c9d97c6480d" [ 1024.425991] env[65107]: _type = "Task" [ 1024.425991] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.435254] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed889b-5ceb-fe52-baee-9c9d97c6480d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.622648] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103332, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.690893] env[65107]: INFO nova.compute.manager [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Took 30.13 seconds to build instance. [ 1024.865297] env[65107]: DEBUG nova.scheduler.client.report [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1024.893696] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103333, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.895435] env[65107]: DEBUG oslo_concurrency.lockutils [req-4f228347-c205-49ad-bb8b-1883af36f51d req-0a5acce9-6b85-4218-8f6d-0f0dba53c3bc service nova] Releasing lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.936686] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ed889b-5ceb-fe52-baee-9c9d97c6480d, 'name': SearchDatastore_Task, 'duration_secs': 0.010058} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.937607] env[65107]: DEBUG oslo_concurrency.lockutils [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.937869] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. {{(pid=65107) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1024.938161] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e233dfcf-c7c1-40cc-af47-bab7965ed782 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.945281] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1024.945281] env[65107]: value = "task-5103334" [ 1024.945281] env[65107]: _type = "Task" [ 1024.945281] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.953573] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103334, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.059462] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee6eb5d-7ca4-4b71-a195-bad97aa3bc8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.067056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Suspending the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1025.067306] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-9eadc131-5716-4333-8559-eafc336485e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.074403] env[65107]: DEBUG oslo_vmware.api [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1025.074403] env[65107]: value = "task-5103335" [ 1025.074403] env[65107]: _type = "Task" [ 1025.074403] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.084387] env[65107]: DEBUG oslo_vmware.api [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103335, 'name': SuspendVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.126566] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103332, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.192682] env[65107]: DEBUG oslo_concurrency.lockutils [None req-66b630c8-0437-497e-a037-4b0ce08b6077 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.646s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.373259] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.999s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.374184] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1025.379160] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.055s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.379482] env[65107]: DEBUG nova.objects.instance [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lazy-loading 'resources' on Instance uuid ac631507-8abf-4019-bdd5-7684dd51d1e9 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.398610] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103333, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517785} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.398933] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65/00574b77-dad6-4f0a-bbcc-20a2b4a5df65.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1025.399226] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.399988] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3138ea4b-fb56-46ac-93c2-7bd86e15b8fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.407861] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1025.412682] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1025.412682] env[65107]: value = "task-5103336" [ 1025.412682] env[65107]: _type = "Task" [ 1025.412682] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.425200] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103336, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.446085] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1025.446085] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1025.446421] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1025.446421] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1025.446589] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1025.446707] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1025.446953] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.447153] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1025.447329] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1025.447509] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1025.447688] env[65107]: DEBUG nova.virt.hardware [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1025.449196] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e81382-a4d1-4d21-90e2-fe7179eb9998 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.462669] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103334, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482647} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.464803] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. [ 1025.466176] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064ad063-faa7-4fac-9cd2-297768842249 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.470178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e70d10-b31c-4028-ba89-42ac436bd7dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.514420] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.514962] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c4efc7b-4d32-4987-9e57-446466e65abf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.538098] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1025.538098] env[65107]: value = "task-5103337" [ 1025.538098] env[65107]: _type = "Task" [ 1025.538098] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.547516] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103337, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.585863] env[65107]: DEBUG oslo_vmware.api [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103335, 'name': SuspendVM_Task} progress is 54%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.624956] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103332, 'name': CreateVM_Task, 'duration_secs': 1.355694} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.625194] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1025.625569] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.625732] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1025.626095] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1025.626332] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baf27ae7-0752-4890-bf52-24d5264a1fa0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.631421] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1025.631421] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202225a-a84e-70f0-e8a4-3d31a9348912" [ 1025.631421] env[65107]: _type = "Task" [ 1025.631421] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.640223] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202225a-a84e-70f0-e8a4-3d31a9348912, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.875957] env[65107]: DEBUG nova.network.neutron [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Successfully updated port: 4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1025.886127] env[65107]: DEBUG nova.compute.utils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1025.890791] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1025.890791] env[65107]: DEBUG nova.network.neutron [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1025.891057] env[65107]: WARNING neutronclient.v2_0.client [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.891366] env[65107]: WARNING neutronclient.v2_0.client [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.891946] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.892294] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.926804] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103336, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084165} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.928177] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.929228] env[65107]: DEBUG nova.compute.manager [req-aafb2df7-4875-4d32-8276-c5c37e3eeb3c req-7afe2404-aaa2-49d3-9e87-779485f33cfc service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Received event network-vif-plugged-4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1025.929555] env[65107]: DEBUG oslo_concurrency.lockutils [req-aafb2df7-4875-4d32-8276-c5c37e3eeb3c req-7afe2404-aaa2-49d3-9e87-779485f33cfc service nova] Acquiring lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.929833] env[65107]: DEBUG oslo_concurrency.lockutils [req-aafb2df7-4875-4d32-8276-c5c37e3eeb3c req-7afe2404-aaa2-49d3-9e87-779485f33cfc service nova] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.930043] env[65107]: DEBUG oslo_concurrency.lockutils [req-aafb2df7-4875-4d32-8276-c5c37e3eeb3c req-7afe2404-aaa2-49d3-9e87-779485f33cfc service nova] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.930224] env[65107]: DEBUG nova.compute.manager [req-aafb2df7-4875-4d32-8276-c5c37e3eeb3c req-7afe2404-aaa2-49d3-9e87-779485f33cfc service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] No waiting events found dispatching network-vif-plugged-4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1025.930445] env[65107]: WARNING nova.compute.manager [req-aafb2df7-4875-4d32-8276-c5c37e3eeb3c req-7afe2404-aaa2-49d3-9e87-779485f33cfc service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Received unexpected event network-vif-plugged-4fe4b3ad-385c-4532-8edc-066c218ced49 for instance with vm_state building and task_state spawning. [ 1025.931385] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9326a1c-6286-4ed4-897d-916b046881c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.960293] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65/00574b77-dad6-4f0a-bbcc-20a2b4a5df65.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.966459] env[65107]: DEBUG nova.policy [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1025.968885] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d22937e-017e-402e-9819-589212de4325 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.994269] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1025.994269] env[65107]: value = "task-5103338" [ 1025.994269] env[65107]: _type = "Task" [ 1025.994269] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.008166] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103338, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.048608] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103337, 'name': ReconfigVM_Task, 'duration_secs': 0.443965} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.051320] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfigured VM instance instance-00000050 to attach disk [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.052394] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13350ff5-1bc5-4bf0-b274-f305d59f57ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.089517] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ab59f16-2381-4113-bc69-edbc2164c8c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.109110] env[65107]: DEBUG oslo_vmware.api [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103335, 'name': SuspendVM_Task, 'duration_secs': 0.792678} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.110414] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Suspended the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1026.110657] env[65107]: DEBUG nova.compute.manager [None req-9a9cbb73-c601-4edd-be91-a291dda7c1f8 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1026.110998] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1026.110998] env[65107]: value = "task-5103339" [ 1026.110998] env[65107]: _type = "Task" [ 1026.110998] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.112304] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c88bcd6-5260-4718-866a-a662198bbcc1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.130580] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103339, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.143049] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5202225a-a84e-70f0-e8a4-3d31a9348912, 'name': SearchDatastore_Task, 'duration_secs': 0.011326} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.143374] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1026.143665] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.143665] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.143849] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.143942] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.144935] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f5f954a-0f21-4587-83e9-b31ed9df92d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.154974] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.154974] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1026.155524] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27b3ef8d-85c6-4b11-b52c-cf61016fd56d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.165101] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1026.165101] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520b571d-b1ed-b213-f4ea-b019392202f4" [ 1026.165101] env[65107]: _type = "Task" [ 1026.165101] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.175013] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "bfe544ec-9b08-4118-a940-a51520ecaac0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.175212] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.175479] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "bfe544ec-9b08-4118-a940-a51520ecaac0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.175713] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.175844] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.178191] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520b571d-b1ed-b213-f4ea-b019392202f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.181900] env[65107]: INFO nova.compute.manager [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Terminating instance [ 1026.326374] env[65107]: DEBUG nova.network.neutron [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Successfully created port: 966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1026.332902] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79909510-43da-4a7b-a3ce-4220b3000bf0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.341347] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660d665a-ab81-4d96-91b6-3a1893a49bc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.374970] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5865356-c217-4f3d-9dec-5ff7939e7dc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.380619] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "refresh_cache-d67468c6-9aaf-49c4-afe4-5d9856c4af7a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.380796] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "refresh_cache-d67468c6-9aaf-49c4-afe4-5d9856c4af7a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.381014] env[65107]: DEBUG nova.network.neutron [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1026.386137] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22cda79-23b0-47b3-831e-4b6263b4cfe6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.391808] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1026.406138] env[65107]: DEBUG nova.compute.provider_tree [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.504459] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103338, 'name': ReconfigVM_Task, 'duration_secs': 0.316565} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.504690] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65/00574b77-dad6-4f0a-bbcc-20a2b4a5df65.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.505366] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba42f530-f92b-40eb-9284-bff2e159850d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.512969] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1026.512969] env[65107]: value = "task-5103340" [ 1026.512969] env[65107]: _type = "Task" [ 1026.512969] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.522783] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103340, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.624529] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103339, 'name': ReconfigVM_Task, 'duration_secs': 0.189452} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.624838] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.625123] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfa917ac-03da-4929-82bf-7a68a6908c60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.636127] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1026.636127] env[65107]: value = "task-5103341" [ 1026.636127] env[65107]: _type = "Task" [ 1026.636127] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.646892] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103341, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.676239] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520b571d-b1ed-b213-f4ea-b019392202f4, 'name': SearchDatastore_Task, 'duration_secs': 0.011328} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.677112] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c9176df-3fc5-45b9-8c31-bb1e16e3207d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.683536] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1026.683536] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5281b92d-dcb5-1dbc-b5c8-ed5191b6cde9" [ 1026.683536] env[65107]: _type = "Task" [ 1026.683536] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.687970] env[65107]: DEBUG nova.compute.manager [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1026.687970] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.689045] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b9e68a-2dad-4b99-9d42-eb416d002c51 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.699206] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5281b92d-dcb5-1dbc-b5c8-ed5191b6cde9, 'name': SearchDatastore_Task, 'duration_secs': 0.010347} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.702025] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1026.702025] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1026.702237] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1026.702304] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f75d121-b872-471e-ab3c-7f5fe8eaac8d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.704268] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d70f71d0-bdd7-4786-9b20-00ce18cdb836 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.711326] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1026.711326] env[65107]: value = "task-5103343" [ 1026.711326] env[65107]: _type = "Task" [ 1026.711326] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.712747] env[65107]: DEBUG oslo_vmware.api [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 1026.712747] env[65107]: value = "task-5103342" [ 1026.712747] env[65107]: _type = "Task" [ 1026.712747] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.726899] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103343, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.732312] env[65107]: DEBUG oslo_vmware.api [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5103342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.886503] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.887051] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.909262] env[65107]: DEBUG nova.scheduler.client.report [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1026.930960] env[65107]: DEBUG nova.network.neutron [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1026.989607] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.990061] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.023567] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103340, 'name': Rename_Task, 'duration_secs': 0.156508} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.023911] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1027.024210] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9135a311-b3b0-48eb-8325-eb92e944f1c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.032055] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1027.032055] env[65107]: value = "task-5103344" [ 1027.032055] env[65107]: _type = "Task" [ 1027.032055] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.041461] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103344, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.140135] env[65107]: WARNING neutronclient.v2_0.client [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.141070] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.141597] env[65107]: WARNING openstack [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.166438] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.229271] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103343, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.233182] env[65107]: DEBUG oslo_vmware.api [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5103342, 'name': PowerOffVM_Task, 'duration_secs': 0.260723} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.233182] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.233182] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.233343] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d24b0963-6a86-46ba-81d7-f1aad5038a9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.318546] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.318903] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.319220] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Deleting the datastore file [datastore1] bfe544ec-9b08-4118-a940-a51520ecaac0 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.319657] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-071aae31-46cf-465a-b39e-5a245a2105e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.329355] env[65107]: DEBUG oslo_vmware.api [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for the task: (returnval){ [ 1027.329355] env[65107]: value = "task-5103346" [ 1027.329355] env[65107]: _type = "Task" [ 1027.329355] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.342642] env[65107]: DEBUG oslo_vmware.api [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5103346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.346470] env[65107]: DEBUG nova.network.neutron [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Updating instance_info_cache with network_info: [{"id": "4fe4b3ad-385c-4532-8edc-066c218ced49", "address": "fa:16:3e:68:a3:e7", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fe4b3ad-38", "ovs_interfaceid": "4fe4b3ad-385c-4532-8edc-066c218ced49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.402340] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1027.415352] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.036s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.418305] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.243s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.418776] env[65107]: DEBUG nova.objects.instance [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lazy-loading 'resources' on Instance uuid 6d5eb77c-c6e8-4ef1-a945-b3485aedd488 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.433893] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1027.434230] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1027.434401] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1027.434608] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1027.434780] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1027.434940] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1027.435198] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.435565] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1027.435565] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1027.435729] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1027.435971] env[65107]: DEBUG nova.virt.hardware [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1027.437275] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b730f642-4ae0-480d-8995-07bfbb34a863 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.448393] env[65107]: INFO nova.scheduler.client.report [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted allocations for instance ac631507-8abf-4019-bdd5-7684dd51d1e9 [ 1027.450551] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3c3d3e-b767-45de-ad0d-94eeff41fdcb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.545767] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103344, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.664144] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103341, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.723989] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103343, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59772} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.724220] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1027.724433] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.724691] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d9df266-d4ec-4a09-bcf1-aeeb492741c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.731754] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1027.731754] env[65107]: value = "task-5103347" [ 1027.731754] env[65107]: _type = "Task" [ 1027.731754] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.742563] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.840728] env[65107]: DEBUG oslo_vmware.api [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Task: {'id': task-5103346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192153} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.844060] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.844060] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.844060] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.844060] env[65107]: INFO nova.compute.manager [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1027.844060] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1027.844060] env[65107]: DEBUG nova.compute.manager [-] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1027.844060] env[65107]: DEBUG nova.network.neutron [-] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1027.844060] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.844060] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.844060] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.851399] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "refresh_cache-d67468c6-9aaf-49c4-afe4-5d9856c4af7a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.851761] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Instance network_info: |[{"id": "4fe4b3ad-385c-4532-8edc-066c218ced49", "address": "fa:16:3e:68:a3:e7", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fe4b3ad-38", "ovs_interfaceid": "4fe4b3ad-385c-4532-8edc-066c218ced49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1027.852280] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:a3:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bf86b133-2b7b-4cab-8f6f-5a0856d34c7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4fe4b3ad-385c-4532-8edc-066c218ced49', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.862238] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1027.863366] env[65107]: DEBUG nova.network.neutron [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Successfully updated port: 966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1027.865284] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1027.865284] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ddbeae7-05c6-485c-865c-947feae4624a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.889510] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.889510] env[65107]: value = "task-5103348" [ 1027.889510] env[65107]: _type = "Task" [ 1027.889510] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.894469] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.896684] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.896926] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.897163] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.897345] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.897514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.900912] env[65107]: INFO nova.compute.manager [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Terminating instance [ 1027.907561] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103348, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.962950] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8e628dfe-cd8f-4c32-a44d-ef207705ffdf tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "ac631507-8abf-4019-bdd5-7684dd51d1e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.674s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.973107] env[65107]: DEBUG nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Received event network-changed-4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1027.973443] env[65107]: DEBUG nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Refreshing instance network info cache due to event network-changed-4fe4b3ad-385c-4532-8edc-066c218ced49. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1027.973814] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Acquiring lock "refresh_cache-d67468c6-9aaf-49c4-afe4-5d9856c4af7a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.974049] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Acquired lock "refresh_cache-d67468c6-9aaf-49c4-afe4-5d9856c4af7a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.974425] env[65107]: DEBUG nova.network.neutron [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Refreshing network info cache for port 4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1028.044339] env[65107]: DEBUG oslo_vmware.api [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103344, 'name': PowerOnVM_Task, 'duration_secs': 0.528893} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.044644] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.044874] env[65107]: INFO nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Took 6.99 seconds to spawn the instance on the hypervisor. [ 1028.045101] env[65107]: DEBUG nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1028.046109] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477250dd-b13c-4ac1-8b04-7459537d8b37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.171641] env[65107]: DEBUG oslo_vmware.api [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103341, 'name': PowerOnVM_Task, 'duration_secs': 1.294175} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.171919] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.175272] env[65107]: DEBUG nova.compute.manager [None req-235537d7-4765-4450-81f6-f3219526c39b tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1028.175510] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2de3b12-dacd-4d9f-90d7-1e7dea7f50a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.227633] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e6498d-07eb-40dd-a50c-d515f32ed5f8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.240640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471104c6-216b-498a-bd8c-9abcdc280df3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.247327] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.279758] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fb522e-bf0d-4b17-b93b-efd2dce79ba9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.288919] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbab94c-88fb-49a4-a4ba-8fb225142769 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.304493] env[65107]: DEBUG nova.compute.provider_tree [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.382405] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-2c45d232-a948-4ad7-80a0-589c2dc91ec1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.382578] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-2c45d232-a948-4ad7-80a0-589c2dc91ec1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.382760] env[65107]: DEBUG nova.network.neutron [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1028.402023] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103348, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.409073] env[65107]: DEBUG nova.compute.manager [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1028.409297] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1028.410477] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b41b69a-88f4-4993-adfe-0edffa5743d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.419603] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1028.419802] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3d5a62e-2039-4873-b788-fb867ede7901 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.485022] env[65107]: WARNING neutronclient.v2_0.client [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.485022] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.485022] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.503295] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1028.503876] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1028.504404] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleting the datastore file [datastore1] b4f7d003-59fb-4ef4-b5f6-234d5154d198 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.505361] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e65d1c2a-56d8-4621-88ef-5e8d4945421e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.519884] env[65107]: DEBUG oslo_vmware.api [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1028.519884] env[65107]: value = "task-5103350" [ 1028.519884] env[65107]: _type = "Task" [ 1028.519884] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.530072] env[65107]: DEBUG oslo_vmware.api [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.573341] env[65107]: INFO nova.compute.manager [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Took 30.32 seconds to build instance. [ 1028.605941] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.606514] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.688493] env[65107]: WARNING neutronclient.v2_0.client [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.688849] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.689384] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.754637] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103347, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.91765} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.754951] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.755910] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ea2d64-f37f-4ca3-8714-cc9b3fc83faa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.780828] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.781184] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3384be8-7d50-4cd7-b6fd-66d105070716 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.802787] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1028.802787] env[65107]: value = "task-5103351" [ 1028.802787] env[65107]: _type = "Task" [ 1028.802787] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.808888] env[65107]: DEBUG nova.network.neutron [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Updated VIF entry in instance network info cache for port 4fe4b3ad-385c-4532-8edc-066c218ced49. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1028.808888] env[65107]: DEBUG nova.network.neutron [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Updating instance_info_cache with network_info: [{"id": "4fe4b3ad-385c-4532-8edc-066c218ced49", "address": "fa:16:3e:68:a3:e7", "network": {"id": "62198ae7-ac69-4bf9-a4b8-3bc32182bec4", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1116282099-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f9188f3e35e4d06a977e2180918c616", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bf86b133-2b7b-4cab-8f6f-5a0856d34c7b", "external-id": "nsx-vlan-transportzone-557", "segmentation_id": 557, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fe4b3ad-38", "ovs_interfaceid": "4fe4b3ad-385c-4532-8edc-066c218ced49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.810010] env[65107]: DEBUG nova.scheduler.client.report [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.819030] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103351, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.886046] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.886194] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.906069] env[65107]: DEBUG nova.network.neutron [-] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.907393] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103348, 'name': CreateVM_Task, 'duration_secs': 0.758945} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.907460] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.912088] env[65107]: WARNING neutronclient.v2_0.client [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.912088] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.912088] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.912088] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1028.912088] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96049061-0e91-45d3-a003-c80e26b35219 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.916980] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1028.916980] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52def280-5f92-c04a-4a11-5db65cdf76df" [ 1028.916980] env[65107]: _type = "Task" [ 1028.916980] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.923878] env[65107]: DEBUG nova.network.neutron [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1028.930758] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52def280-5f92-c04a-4a11-5db65cdf76df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.951083] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.951698] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.031846] env[65107]: DEBUG oslo_vmware.api [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216883} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.032111] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.032270] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.032452] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.032627] env[65107]: INFO nova.compute.manager [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1029.032885] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1029.033135] env[65107]: DEBUG nova.compute.manager [-] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1029.033253] env[65107]: DEBUG nova.network.neutron [-] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1029.033546] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.034237] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.035044] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.046273] env[65107]: WARNING neutronclient.v2_0.client [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.047942] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.048509] env[65107]: WARNING openstack [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.081562] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8829330-2052-445b-af24-917b90917d61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.839s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.131680] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.217047] env[65107]: DEBUG nova.network.neutron [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Updating instance_info_cache with network_info: [{"id": "966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87", "address": "fa:16:3e:d6:c6:15", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap966ebe8f-6f", "ovs_interfaceid": "966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1029.314702] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Releasing lock "refresh_cache-d67468c6-9aaf-49c4-afe4-5d9856c4af7a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.315468] env[65107]: DEBUG nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Received event network-vif-plugged-966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1029.315468] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Acquiring lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.315468] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.315856] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.315856] env[65107]: DEBUG nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] No waiting events found dispatching network-vif-plugged-966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1029.316026] env[65107]: WARNING nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Received unexpected event network-vif-plugged-966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 for instance with vm_state building and task_state spawning. [ 1029.316250] env[65107]: DEBUG nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Received event network-changed-966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1029.316250] env[65107]: DEBUG nova.compute.manager [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Refreshing instance network info cache due to event network-changed-966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1029.316551] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Acquiring lock "refresh_cache-2c45d232-a948-4ad7-80a0-589c2dc91ec1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.317117] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.899s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.320371] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103351, 'name': ReconfigVM_Task, 'duration_secs': 0.500043} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.320371] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.367s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.321943] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Reconfigured VM instance instance-0000005a to attach disk [datastore2] a3162257-ac6b-4468-a573-3611bba69c46/a3162257-ac6b-4468-a573-3611bba69c46.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.322458] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02456c28-2fa2-4388-8dbd-e5ddb6e63f75 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.331467] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1029.331467] env[65107]: value = "task-5103352" [ 1029.331467] env[65107]: _type = "Task" [ 1029.331467] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.341997] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103352, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.348309] env[65107]: INFO nova.scheduler.client.report [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Deleted allocations for instance 6d5eb77c-c6e8-4ef1-a945-b3485aedd488 [ 1029.409497] env[65107]: INFO nova.compute.manager [-] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Took 1.57 seconds to deallocate network for instance. [ 1029.428198] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52def280-5f92-c04a-4a11-5db65cdf76df, 'name': SearchDatastore_Task, 'duration_secs': 0.012735} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.428499] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.428739] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.430242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.430242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.430242] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.430242] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c785060-dc35-4124-a519-b3f9c521fe53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.441615] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.441615] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.441615] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8850778-58d5-401b-a514-de79356bea1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.448852] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1029.448852] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5216924b-cf10-79e4-af85-02fe8c0c89ca" [ 1029.448852] env[65107]: _type = "Task" [ 1029.448852] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.458307] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5216924b-cf10-79e4-af85-02fe8c0c89ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.724610] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-2c45d232-a948-4ad7-80a0-589c2dc91ec1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.725009] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Instance network_info: |[{"id": "966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87", "address": "fa:16:3e:d6:c6:15", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap966ebe8f-6f", "ovs_interfaceid": "966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1029.725428] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.725645] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.725954] env[65107]: DEBUG nova.objects.instance [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'flavor' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.727309] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Acquired lock "refresh_cache-2c45d232-a948-4ad7-80a0-589c2dc91ec1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.727954] env[65107]: DEBUG nova.network.neutron [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Refreshing network info cache for port 966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1029.729993] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:c6:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.737231] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating folder: Project (645feeb0c8524267bff18c5f37f5f011). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1029.738953] env[65107]: WARNING neutronclient.v2_0.client [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.739599] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.740111] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.746646] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d5cc833-e790-4a0d-a95e-5003464f70ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.759745] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created folder: Project (645feeb0c8524267bff18c5f37f5f011) in parent group-v992574. [ 1029.759985] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating folder: Instances. Parent ref: group-v992839. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1029.760284] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bddc45f7-7c18-4960-8c33-9c38ecd6f73f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.769883] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created folder: Instances in parent group-v992839. [ 1029.770154] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1029.770404] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.770558] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-732fcbb2-8832-4d91-a0bd-ce085e4c417c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.792822] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.792822] env[65107]: value = "task-5103355" [ 1029.792822] env[65107]: _type = "Task" [ 1029.792822] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.801938] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103355, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.847864] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103352, 'name': Rename_Task, 'duration_secs': 0.326779} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.848210] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.848472] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1f90dbc-dbc8-4417-b60e-37d30ab0b5b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.855787] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.856230] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.866359] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1029.866359] env[65107]: value = "task-5103356" [ 1029.866359] env[65107]: _type = "Task" [ 1029.866359] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.866800] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a7492f9-2eff-4aec-a2f9-99f291370554 tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "6d5eb77c-c6e8-4ef1-a945-b3485aedd488" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.714s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.877235] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103356, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.890426] env[65107]: DEBUG nova.network.neutron [-] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1029.916783] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.943465] env[65107]: WARNING neutronclient.v2_0.client [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.944161] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.944491] env[65107]: WARNING openstack [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.962486] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5216924b-cf10-79e4-af85-02fe8c0c89ca, 'name': SearchDatastore_Task, 'duration_secs': 0.012478} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.963326] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59fd992e-f2ed-479e-a06b-805d7857173f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.970950] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1029.970950] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525fd061-8271-81f9-bf8f-fd96f46698e0" [ 1029.970950] env[65107]: _type = "Task" [ 1029.970950] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.983501] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525fd061-8271-81f9-bf8f-fd96f46698e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.054893] env[65107]: DEBUG nova.network.neutron [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Updated VIF entry in instance network info cache for port 966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1030.055215] env[65107]: DEBUG nova.network.neutron [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Updating instance_info_cache with network_info: [{"id": "966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87", "address": "fa:16:3e:d6:c6:15", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap966ebe8f-6f", "ovs_interfaceid": "966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.241243] env[65107]: WARNING neutronclient.v2_0.client [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.241796] env[65107]: WARNING openstack [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.242276] env[65107]: WARNING openstack [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.248484] env[65107]: DEBUG nova.objects.instance [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'pci_requests' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.305058] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103355, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.320842] env[65107]: INFO nova.compute.manager [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Unrescuing [ 1030.321152] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.321310] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.321477] env[65107]: DEBUG nova.network.neutron [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1030.365105] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bfe544ec-9b08-4118-a940-a51520ecaac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.365254] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.365365] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ee50d08a-57fc-4c05-96fe-a11fe2708165 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.365486] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 40e17da0-b459-4619-8e9e-18c3093328c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.365606] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance afa1a4cc-aa43-4e10-982f-edaf5cdee584 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.365777] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 888d60d4-bc9c-48af-97f9-434883884e8a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.365957] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e12cb11f-7d6f-47d6-a31e-bf2654174d38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.366115] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance c5fb4b1c-5540-4645-b996-f397729beb23 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1030.366409] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3df65e64-0c45-4707-960e-8f1767e2d011 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.366409] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 59ff1aa6-8d0d-4885-90d3-33d623dca8bb is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1030.366563] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.386008] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103356, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.393826] env[65107]: INFO nova.compute.manager [-] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Took 1.36 seconds to deallocate network for instance. [ 1030.409697] env[65107]: DEBUG nova.compute.manager [req-90e342c3-9853-4afa-bf22-cb1fdbf39ea8 req-7554a051-a746-4b65-9111-66837bc703de service nova] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Received event network-vif-deleted-0d76af9c-1e77-4b0e-908b-551c74fe95ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1030.410016] env[65107]: DEBUG nova.compute.manager [req-90e342c3-9853-4afa-bf22-cb1fdbf39ea8 req-7554a051-a746-4b65-9111-66837bc703de service nova] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Received event network-vif-deleted-d252ac0c-f7eb-4a04-8af4-7b1eb031aa93 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1030.482756] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525fd061-8271-81f9-bf8f-fd96f46698e0, 'name': SearchDatastore_Task, 'duration_secs': 0.011592} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.482957] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.483216] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] d67468c6-9aaf-49c4-afe4-5d9856c4af7a/d67468c6-9aaf-49c4-afe4-5d9856c4af7a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.483482] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bb0246c-49f5-40ef-be62-edc668ab544f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.490361] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1030.490361] env[65107]: value = "task-5103357" [ 1030.490361] env[65107]: _type = "Task" [ 1030.490361] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.499540] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.558164] env[65107]: DEBUG oslo_concurrency.lockutils [req-ec09c811-c520-436f-a97d-854686c6e82a req-4387147a-2d59-48d5-99c7-554cd91493e2 service nova] Releasing lock "refresh_cache-2c45d232-a948-4ad7-80a0-589c2dc91ec1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.752507] env[65107]: DEBUG nova.objects.base [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1030.752800] env[65107]: DEBUG nova.network.neutron [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1030.753082] env[65107]: WARNING neutronclient.v2_0.client [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.753494] env[65107]: WARNING neutronclient.v2_0.client [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.754154] env[65107]: WARNING openstack [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.754545] env[65107]: WARNING openstack [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.806969] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103355, 'name': CreateVM_Task, 'duration_secs': 0.984923} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.812359] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.813570] env[65107]: WARNING neutronclient.v2_0.client [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.813808] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.813963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.814366] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1030.815048] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7a2e080-dcf7-417d-ad2c-8570cf4a2a53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.823505] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1030.823505] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529027b5-05de-41f7-6000-fa2e74329a83" [ 1030.823505] env[65107]: _type = "Task" [ 1030.823505] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.824133] env[65107]: WARNING neutronclient.v2_0.client [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.824761] env[65107]: WARNING openstack [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.825154] env[65107]: WARNING openstack [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.843597] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529027b5-05de-41f7-6000-fa2e74329a83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.851883] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0ee49b73-0093-457e-9b4c-b0bf0339d610 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.126s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.870139] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d2e7eee7-c9b2-49aa-92f9-1efd7c901509 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1030.870344] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance a3162257-ac6b-4468-a573-3611bba69c46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.870461] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b4f7d003-59fb-4ef4-b5f6-234d5154d198 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.870564] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 00574b77-dad6-4f0a-bbcc-20a2b4a5df65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.870676] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d67468c6-9aaf-49c4-afe4-5d9856c4af7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.870791] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 2c45d232-a948-4ad7-80a0-589c2dc91ec1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1030.882198] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "888d60d4-bc9c-48af-97f9-434883884e8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.882577] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "888d60d4-bc9c-48af-97f9-434883884e8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.882853] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "888d60d4-bc9c-48af-97f9-434883884e8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.883053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "888d60d4-bc9c-48af-97f9-434883884e8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.883296] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "888d60d4-bc9c-48af-97f9-434883884e8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.887377] env[65107]: INFO nova.compute.manager [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Terminating instance [ 1030.894360] env[65107]: DEBUG oslo_vmware.api [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103356, 'name': PowerOnVM_Task, 'duration_secs': 0.57826} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.894927] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.895168] env[65107]: DEBUG nova.compute.manager [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1030.896239] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e9262c-06b8-44b6-bd3b-c2fbb6301320 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.901536] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.972858] env[65107]: WARNING openstack [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.973367] env[65107]: WARNING openstack [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.002351] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103357, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493682} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.006948] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] d67468c6-9aaf-49c4-afe4-5d9856c4af7a/d67468c6-9aaf-49c4-afe4-5d9856c4af7a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.007190] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.007654] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17bf6f21-b15c-4db0-b6f4-5f3967791d51 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.019528] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1031.019528] env[65107]: value = "task-5103358" [ 1031.019528] env[65107]: _type = "Task" [ 1031.019528] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.030183] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103358, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.037937] env[65107]: WARNING neutronclient.v2_0.client [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.038733] env[65107]: WARNING openstack [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.039013] env[65107]: WARNING openstack [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.133872] env[65107]: DEBUG nova.network.neutron [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.335219] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529027b5-05de-41f7-6000-fa2e74329a83, 'name': SearchDatastore_Task, 'duration_secs': 0.060481} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.335607] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.335726] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.335965] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.336217] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.336635] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.336727] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c53b43a-b6cf-4ca9-9386-8f87da663d28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.346129] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.346344] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.347179] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca236ae5-6c3f-469b-afc9-afed97b0c44d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.353283] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1031.353283] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52acb7d2-2cca-f982-aa3e-d8251ff0b979" [ 1031.353283] env[65107]: _type = "Task" [ 1031.353283] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.363084] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52acb7d2-2cca-f982-aa3e-d8251ff0b979, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.384063] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 04a508b4-4430-411c-8573-726065b558a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1031.384063] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1031.384063] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=100GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '14', 'num_vm_active': '9', 'num_task_deleting': '2', 'num_os_type_None': '14', 'num_proj_e1fd7913632a4a5b9e076b5b156fa5fa': '1', 'io_workload': '2', 'num_task_None': '9', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '1', 'num_proj_bfb7fca6907c4c32a1aa6b8023c39afb': '3', 'num_vm_rescued': '2', 'num_proj_152868ee0723449baac10f6a8ceeb738': '1', 'num_proj_4daf9bbfa3e9477d83d9d8c3d1fb5a02': '1', 'num_proj_7784762e36ae4c6fb30012ec801c3b88': '1', 'num_proj_f379144b78764fe394039d87b043a946': '1', 'num_task_rebuild_spawning': '1', 'num_proj_8ff69cbe020d4f28a1374e9ce3721416': '1', 'num_vm_suspended': '1', 'num_proj_4778fe9152224fd29d1f6220a19b5a36': '1', 'num_proj_c805fe68d0fd4147b9b4b7fdc9d6a6d0': '1', 'num_vm_building': '2', 'num_task_spawning': '2', 'num_proj_6f9188f3e35e4d06a977e2180918c616': '1', 'num_proj_645feeb0c8524267bff18c5f37f5f011': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1031.395703] env[65107]: DEBUG nova.compute.manager [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1031.395703] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.396629] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7d9a00-33d2-457f-aa3d-84720e3a3f23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.406849] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.407163] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f436450-711e-4ce6-beaf-0d23c678255e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.417833] env[65107]: DEBUG oslo_vmware.api [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 1031.417833] env[65107]: value = "task-5103359" [ 1031.417833] env[65107]: _type = "Task" [ 1031.417833] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.421455] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.430441] env[65107]: DEBUG oslo_vmware.api [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.531960] env[65107]: DEBUG nova.compute.manager [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1031.542085] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103358, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073238} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.542085] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.542085] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf9f218-c0a8-4a86-89a6-80bba197a6ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.565462] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] d67468c6-9aaf-49c4-afe4-5d9856c4af7a/d67468c6-9aaf-49c4-afe4-5d9856c4af7a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.568826] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e736e6d4-1acd-4092-a03f-be962a66aa6d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.597979] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1031.597979] env[65107]: value = "task-5103360" [ 1031.597979] env[65107]: _type = "Task" [ 1031.597979] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.613653] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103360, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.637218] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.638033] env[65107]: DEBUG nova.objects.instance [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'flavor' on Instance uuid e12cb11f-7d6f-47d6-a31e-bf2654174d38 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.709386] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e97204-006f-4bc1-a982-767e52564276 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.717183] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fda9f3-0100-4047-b3cb-b418474d0f48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.748188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153c38ff-12d7-4803-90f2-853a54311a42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.756665] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7980c3-ce82-4a5f-846e-4d41ce6d017d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.771653] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.866594] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52acb7d2-2cca-f982-aa3e-d8251ff0b979, 'name': SearchDatastore_Task, 'duration_secs': 0.010431} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.868792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "a3162257-ac6b-4468-a573-3611bba69c46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.868792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "a3162257-ac6b-4468-a573-3611bba69c46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.869068] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "a3162257-ac6b-4468-a573-3611bba69c46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.869258] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "a3162257-ac6b-4468-a573-3611bba69c46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.869603] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "a3162257-ac6b-4468-a573-3611bba69c46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.871468] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfe04db5-8956-4dc3-a06d-cbd22bdc52af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.875204] env[65107]: INFO nova.compute.manager [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Terminating instance [ 1031.879832] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1031.879832] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b788be-0881-542b-2195-9e767530ab30" [ 1031.879832] env[65107]: _type = "Task" [ 1031.879832] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.890446] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b788be-0881-542b-2195-9e767530ab30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.931477] env[65107]: DEBUG oslo_vmware.api [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103359, 'name': PowerOffVM_Task, 'duration_secs': 0.207632} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.932435] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.932651] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1031.932913] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f88a32d2-7264-412c-90c9-3ddc82829c99 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.007771] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.008212] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.008529] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Deleting the datastore file [datastore1] 888d60d4-bc9c-48af-97f9-434883884e8a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.008925] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66b1c081-db03-4a14-a137-d800eb0ce980 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.018055] env[65107]: DEBUG oslo_vmware.api [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for the task: (returnval){ [ 1032.018055] env[65107]: value = "task-5103362" [ 1032.018055] env[65107]: _type = "Task" [ 1032.018055] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.027421] env[65107]: DEBUG oslo_vmware.api [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103362, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.055226] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.111978] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103360, 'name': ReconfigVM_Task, 'duration_secs': 0.314373} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.112340] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Reconfigured VM instance instance-0000005d to attach disk [datastore1] d67468c6-9aaf-49c4-afe4-5d9856c4af7a/d67468c6-9aaf-49c4-afe4-5d9856c4af7a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.112998] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b9f628a-d341-4ddc-856f-5d4fc9c0717e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.120611] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1032.120611] env[65107]: value = "task-5103363" [ 1032.120611] env[65107]: _type = "Task" [ 1032.120611] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.136928] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103363, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.144648] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18b00bd-e002-48e4-92ce-ce07108d56d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.170151] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.170573] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52bca3b8-bc84-4119-a347-a2c016e6ef80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.178612] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1032.178612] env[65107]: value = "task-5103364" [ 1032.178612] env[65107]: _type = "Task" [ 1032.178612] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.187802] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.275595] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1032.379150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "refresh_cache-a3162257-ac6b-4468-a573-3611bba69c46" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.379150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquired lock "refresh_cache-a3162257-ac6b-4468-a573-3611bba69c46" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.379150] env[65107]: DEBUG nova.network.neutron [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1032.391720] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b788be-0881-542b-2195-9e767530ab30, 'name': SearchDatastore_Task, 'duration_secs': 0.012236} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.391992] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.392268] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2c45d232-a948-4ad7-80a0-589c2dc91ec1/2c45d232-a948-4ad7-80a0-589c2dc91ec1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.392529] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a36f8c2d-46ff-454d-ad62-82c07094fdfa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.400841] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1032.400841] env[65107]: value = "task-5103365" [ 1032.400841] env[65107]: _type = "Task" [ 1032.400841] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.412144] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.528321] env[65107]: DEBUG oslo_vmware.api [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Task: {'id': task-5103362, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157942} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.528588] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.528823] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1032.529080] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.529265] env[65107]: INFO nova.compute.manager [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1032.529513] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1032.529709] env[65107]: DEBUG nova.compute.manager [-] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1032.529808] env[65107]: DEBUG nova.network.neutron [-] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1032.530074] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.530596] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.530892] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.570463] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.631452] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103363, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.691997] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103364, 'name': PowerOffVM_Task, 'duration_secs': 0.244447} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.692341] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.697666] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfiguring VM instance instance-00000050 to detach disk 2002 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1032.697929] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46a55ccf-ac09-4493-a8f5-374e347131ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.720549] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1032.720549] env[65107]: value = "task-5103366" [ 1032.720549] env[65107]: _type = "Task" [ 1032.720549] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.730662] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.782222] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1032.782222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.462s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.782561] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.725s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.782617] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.787920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.867s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.788570] env[65107]: INFO nova.compute.claims [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1032.817318] env[65107]: INFO nova.scheduler.client.report [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Deleted allocations for instance 59ff1aa6-8d0d-4885-90d3-33d623dca8bb [ 1032.828301] env[65107]: DEBUG nova.compute.manager [req-64f15508-86a0-482b-929e-fd41c8bd6510 req-c8edd40d-0296-4c1d-950e-d5f88e6fcdd9 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Received event network-vif-deleted-dacbb269-1799-4b10-8007-7778dddae890 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1032.828571] env[65107]: INFO nova.compute.manager [req-64f15508-86a0-482b-929e-fd41c8bd6510 req-c8edd40d-0296-4c1d-950e-d5f88e6fcdd9 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Neutron deleted interface dacbb269-1799-4b10-8007-7778dddae890; detaching it from the instance and deleting it from the info cache [ 1032.828711] env[65107]: DEBUG nova.network.neutron [req-64f15508-86a0-482b-929e-fd41c8bd6510 req-c8edd40d-0296-4c1d-950e-d5f88e6fcdd9 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1032.882472] env[65107]: WARNING neutronclient.v2_0.client [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.882472] env[65107]: WARNING openstack [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.882874] env[65107]: WARNING openstack [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.912606] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.913730] env[65107]: DEBUG nova.network.neutron [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1032.933591] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.933981] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.934399] env[65107]: DEBUG nova.objects.instance [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'flavor' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.976801] env[65107]: DEBUG nova.network.neutron [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1033.138882] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103363, 'name': Rename_Task, 'duration_secs': 0.948337} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.141985] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.144020] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-281a8a7d-7ab6-429f-8e51-83bd77b902c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.152142] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1033.152142] env[65107]: value = "task-5103367" [ 1033.152142] env[65107]: _type = "Task" [ 1033.152142] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.163126] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.233554] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103366, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.301234] env[65107]: DEBUG nova.network.neutron [-] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1033.303723] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.303723] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.305229] env[65107]: INFO nova.compute.manager [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Rebooting instance [ 1033.326160] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9adc7b05-f4b6-4c1e-b322-c9e0d4a75c62 tempest-ServerAddressesNegativeTestJSON-1826277776 tempest-ServerAddressesNegativeTestJSON-1826277776-project-member] Lock "59ff1aa6-8d0d-4885-90d3-33d623dca8bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.030s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.332209] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abc9e1df-2686-4b12-9e8a-de29f299f5df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.343956] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96169a8c-dd39-4dca-b46b-6c14e1d16852 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.382708] env[65107]: DEBUG nova.compute.manager [req-64f15508-86a0-482b-929e-fd41c8bd6510 req-c8edd40d-0296-4c1d-950e-d5f88e6fcdd9 service nova] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Detach interface failed, port_id=dacbb269-1799-4b10-8007-7778dddae890, reason: Instance 888d60d4-bc9c-48af-97f9-434883884e8a could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1033.411012] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.948323} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.411290] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2c45d232-a948-4ad7-80a0-589c2dc91ec1/2c45d232-a948-4ad7-80a0-589c2dc91ec1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.411540] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.411803] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3bf8ca4f-8644-45b9-a9bb-f926655dc284 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.419521] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1033.419521] env[65107]: value = "task-5103368" [ 1033.419521] env[65107]: _type = "Task" [ 1033.419521] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.430340] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103368, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.439050] env[65107]: WARNING neutronclient.v2_0.client [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1033.439890] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1033.440572] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1033.480042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Releasing lock "refresh_cache-a3162257-ac6b-4468-a573-3611bba69c46" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1033.481060] env[65107]: DEBUG nova.compute.manager [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1033.481334] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.482657] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bffb74-9753-4608-ae4c-71a6b59c33ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.495435] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.495738] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-327c2977-f6b4-4e06-b6c1-2da2a16b4adb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.503822] env[65107]: DEBUG oslo_vmware.api [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1033.503822] env[65107]: value = "task-5103369" [ 1033.503822] env[65107]: _type = "Task" [ 1033.503822] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.513568] env[65107]: DEBUG oslo_vmware.api [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.530911] env[65107]: DEBUG nova.objects.instance [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'pci_requests' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.663471] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103367, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.732516] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103366, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.806655] env[65107]: INFO nova.compute.manager [-] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Took 1.28 seconds to deallocate network for instance. [ 1033.825836] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.826126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1033.826940] env[65107]: DEBUG nova.network.neutron [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1033.935899] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103368, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070032} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.940436] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.942165] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265d2868-1530-42a8-8479-b93d3130f1bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.981852] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 2c45d232-a948-4ad7-80a0-589c2dc91ec1/2c45d232-a948-4ad7-80a0-589c2dc91ec1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.986662] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2354271-0d36-402e-828f-ef832dd0bae8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.028025] env[65107]: DEBUG oslo_vmware.api [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103369, 'name': PowerOffVM_Task, 'duration_secs': 0.509356} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.030225] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.030520] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1034.031266] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1034.031266] env[65107]: value = "task-5103370" [ 1034.031266] env[65107]: _type = "Task" [ 1034.031266] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.035204] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2fadafec-f524-4048-8819-07e84011a16d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.038268] env[65107]: DEBUG nova.objects.base [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1034.038612] env[65107]: DEBUG nova.network.neutron [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1034.039172] env[65107]: WARNING neutronclient.v2_0.client [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.039640] env[65107]: WARNING neutronclient.v2_0.client [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.040569] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.041137] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.067622] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.086644] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1034.086876] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1034.087075] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Deleting the datastore file [datastore2] a3162257-ac6b-4468-a573-3611bba69c46 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.087358] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53a8c96a-462c-4d17-b584-a44cbd90a2e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.094302] env[65107]: DEBUG oslo_vmware.api [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for the task: (returnval){ [ 1034.094302] env[65107]: value = "task-5103372" [ 1034.094302] env[65107]: _type = "Task" [ 1034.094302] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.100138] env[65107]: DEBUG nova.policy [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1034.110750] env[65107]: DEBUG oslo_vmware.api [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103372, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.166263] env[65107]: DEBUG oslo_vmware.api [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103367, 'name': PowerOnVM_Task, 'duration_secs': 0.800939} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.166600] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.166809] env[65107]: INFO nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Took 8.76 seconds to spawn the instance on the hypervisor. [ 1034.166993] env[65107]: DEBUG nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1034.167872] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ae3cde-645b-49ef-812a-4fdb4dd5f725 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.186787] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3a24ce-f0f5-48d6-a1d9-48b57e959e22 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.195253] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b052f656-3a9a-466c-8f6c-c3c107147cff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.232978] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4032dcf9-54e7-4f9f-bfed-f5940d82cb3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.245219] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a172f210-1993-4709-a97f-6623d16a3a14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.250131] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103366, 'name': ReconfigVM_Task, 'duration_secs': 1.27716} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.250489] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfigured VM instance instance-00000050 to detach disk 2002 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1034.250754] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.251450] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-213d64c7-58ec-40ee-9999-80c9834a3632 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.263767] env[65107]: DEBUG nova.compute.provider_tree [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.270702] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1034.270702] env[65107]: value = "task-5103373" [ 1034.270702] env[65107]: _type = "Task" [ 1034.270702] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.280360] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.316556] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.329581] env[65107]: WARNING neutronclient.v2_0.client [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.330681] env[65107]: WARNING openstack [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.331075] env[65107]: WARNING openstack [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.467425] env[65107]: DEBUG nova.network.neutron [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Successfully created port: fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1034.548849] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.609286] env[65107]: DEBUG oslo_vmware.api [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Task: {'id': task-5103372, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299813} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.609570] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.609766] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.609983] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.610216] env[65107]: INFO nova.compute.manager [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1034.610475] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1034.610690] env[65107]: DEBUG nova.compute.manager [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1034.610815] env[65107]: DEBUG nova.network.neutron [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1034.611067] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.611616] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.611874] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.690863] env[65107]: INFO nova.compute.manager [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Took 32.57 seconds to build instance. [ 1034.768749] env[65107]: DEBUG nova.scheduler.client.report [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1034.783609] env[65107]: DEBUG oslo_vmware.api [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103373, 'name': PowerOnVM_Task, 'duration_secs': 0.396973} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.783724] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.784012] env[65107]: DEBUG nova.compute.manager [None req-04ad341f-5d89-466d-8ed3-e0a47e7c7872 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1034.784847] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f028045-9f52-476d-9ea2-c0c926125555 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.838124] env[65107]: DEBUG nova.network.neutron [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1034.838545] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.051837] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103370, 'name': ReconfigVM_Task, 'duration_secs': 0.877114} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.051837] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 2c45d232-a948-4ad7-80a0-589c2dc91ec1/2c45d232-a948-4ad7-80a0-589c2dc91ec1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.051837] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3bb4ea54-9039-4536-b64e-1fc47d829abf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.058409] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1035.058409] env[65107]: value = "task-5103374" [ 1035.058409] env[65107]: _type = "Task" [ 1035.058409] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.068430] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103374, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.192917] env[65107]: DEBUG oslo_concurrency.lockutils [None req-07682a2f-9adb-4a6f-b705-3ba5dbdc2a94 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.080s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.244368] env[65107]: WARNING openstack [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.244851] env[65107]: WARNING openstack [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.275150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.275906] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1035.282548] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.789s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.282841] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.287477] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.110s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.287791] env[65107]: DEBUG nova.objects.instance [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'pci_requests' on Instance uuid d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.327589] env[65107]: INFO nova.scheduler.client.report [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Deleted allocations for instance c5fb4b1c-5540-4645-b996-f397729beb23 [ 1035.341996] env[65107]: DEBUG nova.network.neutron [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1035.568888] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103374, 'name': Rename_Task, 'duration_secs': 0.201385} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.569305] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.569449] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-389057b9-1acf-4f20-8bb0-6665b7be1b06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.577876] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1035.577876] env[65107]: value = "task-5103375" [ 1035.577876] env[65107]: _type = "Task" [ 1035.577876] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.594341] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.731340] env[65107]: WARNING neutronclient.v2_0.client [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.732626] env[65107]: WARNING openstack [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.732626] env[65107]: WARNING openstack [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.792215] env[65107]: DEBUG nova.compute.utils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1035.794991] env[65107]: DEBUG nova.objects.instance [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'numa_topology' on Instance uuid d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.800657] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1035.800877] env[65107]: DEBUG nova.network.neutron [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1035.801645] env[65107]: WARNING neutronclient.v2_0.client [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.801645] env[65107]: WARNING neutronclient.v2_0.client [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.802090] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.802449] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.810627] env[65107]: INFO nova.compute.claims [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.813328] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1035.843405] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cfcaa43c-984f-4d9b-bd7a-af1e960903ad tempest-ServerShowV247Test-1230004577 tempest-ServerShowV247Test-1230004577-project-member] Lock "c5fb4b1c-5540-4645-b996-f397729beb23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.761s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.844935] env[65107]: INFO nova.compute.manager [-] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Took 1.23 seconds to deallocate network for instance. [ 1035.913339] env[65107]: DEBUG nova.policy [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd87f09a43c049deb1fda21602b55a64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '544c649453844164bac53c98b342543f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1035.944774] env[65107]: DEBUG nova.network.neutron [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1036.067930] env[65107]: DEBUG nova.network.neutron [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Successfully updated port: fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1036.098467] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103375, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.351860] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.448018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.574295] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.574611] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.574763] env[65107]: DEBUG nova.network.neutron [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1036.594020] env[65107]: DEBUG oslo_vmware.api [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103375, 'name': PowerOnVM_Task, 'duration_secs': 0.534255} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.594242] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.594455] env[65107]: INFO nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Took 9.19 seconds to spawn the instance on the hypervisor. [ 1036.594644] env[65107]: DEBUG nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1036.595556] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc3ff04-0e0e-458c-b729-2b24b8c9b589 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.835427] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1036.844689] env[65107]: DEBUG nova.network.neutron [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Successfully created port: 96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1036.865391] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1036.866247] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1036.866247] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1036.866247] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1036.866247] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1036.866489] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1036.866568] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.866745] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1036.866921] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1036.867105] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1036.867313] env[65107]: DEBUG nova.virt.hardware [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1036.868325] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff0b0d9-f595-436b-bd0d-14787530815d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.880020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfdf10d-ca1b-46aa-97fb-438bf04e50d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.956319] env[65107]: DEBUG nova.compute.manager [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1036.957298] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53758393-55fa-4267-83a3-fd3267019ac4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.984434] env[65107]: DEBUG nova.compute.manager [req-e005de4a-ec04-4a98-937b-15bb34fdac09 req-ab30f27f-056d-45a9-9026-794c523029c6 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-vif-plugged-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1036.984700] env[65107]: DEBUG oslo_concurrency.lockutils [req-e005de4a-ec04-4a98-937b-15bb34fdac09 req-ab30f27f-056d-45a9-9026-794c523029c6 service nova] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.984971] env[65107]: DEBUG oslo_concurrency.lockutils [req-e005de4a-ec04-4a98-937b-15bb34fdac09 req-ab30f27f-056d-45a9-9026-794c523029c6 service nova] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.989334] env[65107]: DEBUG oslo_concurrency.lockutils [req-e005de4a-ec04-4a98-937b-15bb34fdac09 req-ab30f27f-056d-45a9-9026-794c523029c6 service nova] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.004s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.989607] env[65107]: DEBUG nova.compute.manager [req-e005de4a-ec04-4a98-937b-15bb34fdac09 req-ab30f27f-056d-45a9-9026-794c523029c6 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] No waiting events found dispatching network-vif-plugged-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1036.989822] env[65107]: WARNING nova.compute.manager [req-e005de4a-ec04-4a98-937b-15bb34fdac09 req-ab30f27f-056d-45a9-9026-794c523029c6 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received unexpected event network-vif-plugged-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 for instance with vm_state active and task_state None. [ 1037.078866] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.079311] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.113218] env[65107]: INFO nova.compute.manager [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Took 34.86 seconds to build instance. [ 1037.126418] env[65107]: WARNING nova.network.neutron [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] 40a76aa5-ac46-4067-a98d-cd60410bf479 already exists in list: networks containing: ['40a76aa5-ac46-4067-a98d-cd60410bf479']. ignoring it [ 1037.147747] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fc68a4-d967-4b9c-905c-4bd702c8b738 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.159751] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f269781-ee10-460f-bf5b-0b6544ebe427 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.198040] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.198498] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.205747] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c87a2b-bc74-4c35-9553-6ecd3375477c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.214521] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05128d4-e295-4472-8c66-05ca676c4426 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.231045] env[65107]: DEBUG nova.compute.provider_tree [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.306363] env[65107]: WARNING neutronclient.v2_0.client [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.307194] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.307664] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.424269] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.424663] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.618292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-69aca552-3f3a-4da8-b787-b73aebd1dc66 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.377s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.729817] env[65107]: WARNING neutronclient.v2_0.client [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.730533] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.730887] env[65107]: WARNING openstack [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.739071] env[65107]: DEBUG nova.scheduler.client.report [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1037.833039] env[65107]: DEBUG nova.network.neutron [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "address": "fa:16:3e:6b:22:2f", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbe601d0-6b", "ovs_interfaceid": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1037.975218] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda7b713-6f3c-4fcc-91ea-3833b9ad37de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.986130] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Doing hard reboot of VM {{(pid=65107) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1037.986436] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-74365618-3a78-4f77-9841-2385c10025dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.996178] env[65107]: DEBUG oslo_vmware.api [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1037.996178] env[65107]: value = "task-5103376" [ 1037.996178] env[65107]: _type = "Task" [ 1037.996178] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.010130] env[65107]: DEBUG oslo_vmware.api [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103376, 'name': ResetVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.244396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.957s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.244765] env[65107]: WARNING neutronclient.v2_0.client [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.248139] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.331s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.248378] env[65107]: DEBUG nova.objects.instance [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lazy-loading 'resources' on Instance uuid bfe544ec-9b08-4118-a940-a51520ecaac0 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.336174] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.337039] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.337039] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.338018] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ae2fb0-a612-4cf4-a086-596884e4e605 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.356701] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1038.356937] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1038.357104] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1038.357290] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1038.357495] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1038.357667] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1038.357890] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.358069] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1038.358236] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1038.358417] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1038.358621] env[65107]: DEBUG nova.virt.hardware [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1038.365228] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfiguring VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1038.365582] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-641661d8-7967-4a45-ac5a-cff93e7c6ad1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.383278] env[65107]: DEBUG oslo_vmware.api [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1038.383278] env[65107]: value = "task-5103377" [ 1038.383278] env[65107]: _type = "Task" [ 1038.383278] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.392156] env[65107]: DEBUG oslo_vmware.api [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103377, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.461124] env[65107]: DEBUG nova.network.neutron [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Successfully updated port: 96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1038.511370] env[65107]: DEBUG oslo_vmware.api [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103376, 'name': ResetVM_Task, 'duration_secs': 0.115484} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.512533] env[65107]: INFO nova.network.neutron [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating port b6bba8b1-799c-4c2b-bd94-8726094948ee with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1038.515432] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Did hard reboot of VM {{(pid=65107) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1038.515432] env[65107]: DEBUG nova.compute.manager [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1038.516406] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177b565e-37fc-453c-b6d3-43ee2a685bcf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.576588] env[65107]: DEBUG nova.compute.manager [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1038.578866] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5445f9f-2161-42f5-a6cd-05ca90c2a462 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.894948] env[65107]: DEBUG oslo_vmware.api [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103377, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.965108] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.965108] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.965108] env[65107]: DEBUG nova.network.neutron [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1039.008469] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8837f2-cb96-4b5e-8cd1-4160f7adecbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.017545] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7408c6-47c8-40c7-aaca-3c19bb7d04e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.060894] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1601aae5-518a-4c24-89d5-20e732d2d3b4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.757s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1039.062529] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da12c25-a3a0-4215-90ec-61e43a043f05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.073118] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fbb9e6-796b-4bb9-a3cc-ce40290acef3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.090798] env[65107]: DEBUG nova.compute.provider_tree [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.094126] env[65107]: INFO nova.compute.manager [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] instance snapshotting [ 1039.098811] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d24e32-8ebd-415f-bf22-91cbb7510479 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.122916] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27cd712-1b0f-4620-a431-fa9a8063d064 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.312806] env[65107]: DEBUG nova.compute.manager [req-bb8a4d19-b5e6-42bc-884d-f60fc69d9fe8 req-68128f3f-2db0-4dbd-9904-cd9462a7bc5d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Received event network-vif-plugged-96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1039.313068] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb8a4d19-b5e6-42bc-884d-f60fc69d9fe8 req-68128f3f-2db0-4dbd-9904-cd9462a7bc5d service nova] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1039.313963] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb8a4d19-b5e6-42bc-884d-f60fc69d9fe8 req-68128f3f-2db0-4dbd-9904-cd9462a7bc5d service nova] Lock "04a508b4-4430-411c-8573-726065b558a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1039.313963] env[65107]: DEBUG oslo_concurrency.lockutils [req-bb8a4d19-b5e6-42bc-884d-f60fc69d9fe8 req-68128f3f-2db0-4dbd-9904-cd9462a7bc5d service nova] Lock "04a508b4-4430-411c-8573-726065b558a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1039.313963] env[65107]: DEBUG nova.compute.manager [req-bb8a4d19-b5e6-42bc-884d-f60fc69d9fe8 req-68128f3f-2db0-4dbd-9904-cd9462a7bc5d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] No waiting events found dispatching network-vif-plugged-96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1039.313963] env[65107]: WARNING nova.compute.manager [req-bb8a4d19-b5e6-42bc-884d-f60fc69d9fe8 req-68128f3f-2db0-4dbd-9904-cd9462a7bc5d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Received unexpected event network-vif-plugged-96111c75-74f7-4496-9470-0972fef27f6f for instance with vm_state building and task_state spawning. [ 1039.395147] env[65107]: DEBUG oslo_vmware.api [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103377, 'name': ReconfigVM_Task, 'duration_secs': 0.770236} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.395762] env[65107]: WARNING neutronclient.v2_0.client [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.395864] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1039.396125] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfigured VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1039.470813] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1039.471275] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1039.520054] env[65107]: DEBUG nova.network.neutron [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1039.577282] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1039.577282] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1039.597782] env[65107]: DEBUG nova.scheduler.client.report [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1039.638357] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1039.638468] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9f9ac5a3-a647-4e56-bed5-4ee3452bab25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.647027] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1039.647027] env[65107]: value = "task-5103378" [ 1039.647027] env[65107]: _type = "Task" [ 1039.647027] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.656259] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103378, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.660833] env[65107]: WARNING neutronclient.v2_0.client [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.662036] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1039.662465] env[65107]: WARNING openstack [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1039.796098] env[65107]: DEBUG nova.network.neutron [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updating instance_info_cache with network_info: [{"id": "96111c75-74f7-4496-9470-0972fef27f6f", "address": "fa:16:3e:ab:6f:69", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96111c75-74", "ovs_interfaceid": "96111c75-74f7-4496-9470-0972fef27f6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1039.901657] env[65107]: DEBUG oslo_concurrency.lockutils [None req-04a81454-49ed-4a23-90a9-6c65da3b350b tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.967s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1039.958722] env[65107]: DEBUG nova.compute.manager [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-changed-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1039.958722] env[65107]: DEBUG nova.compute.manager [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing instance network info cache due to event network-changed-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1039.958722] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.958722] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1039.958722] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing network info cache for port fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1040.084941] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.085091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1040.085277] env[65107]: DEBUG nova.network.neutron [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1040.104084] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.107715] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.205s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.107715] env[65107]: DEBUG nova.objects.instance [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'resources' on Instance uuid b4f7d003-59fb-4ef4-b5f6-234d5154d198 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.122961] env[65107]: INFO nova.scheduler.client.report [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Deleted allocations for instance bfe544ec-9b08-4118-a940-a51520ecaac0 [ 1040.162370] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103378, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.299069] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1040.299380] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Instance network_info: |[{"id": "96111c75-74f7-4496-9470-0972fef27f6f", "address": "fa:16:3e:ab:6f:69", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96111c75-74", "ovs_interfaceid": "96111c75-74f7-4496-9470-0972fef27f6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1040.299893] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:6f:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96111c75-74f7-4496-9470-0972fef27f6f', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.307761] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1040.308020] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04a508b4-4430-411c-8573-726065b558a1] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1040.308257] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48bf8a7b-36c6-4878-b0e9-898ac5300730 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.330928] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.330928] env[65107]: value = "task-5103379" [ 1040.330928] env[65107]: _type = "Task" [ 1040.330928] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.341331] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103379, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.462395] env[65107]: WARNING neutronclient.v2_0.client [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.463556] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.464116] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.589429] env[65107]: WARNING neutronclient.v2_0.client [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.589429] env[65107]: WARNING openstack [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.589849] env[65107]: WARNING openstack [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.600143] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.600143] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.647024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a5272568-fdce-4582-8845-5f4e59b4656c tempest-ServersV294TestFqdnHostnames-1427429126 tempest-ServersV294TestFqdnHostnames-1427429126-project-member] Lock "bfe544ec-9b08-4118-a940-a51520ecaac0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.469s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.657799] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103378, 'name': CreateSnapshot_Task, 'duration_secs': 0.812711} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.658093] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1040.658952] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5b2b3d-f3a8-4156-9b8d-c386dec89435 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.704676] env[65107]: WARNING neutronclient.v2_0.client [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.705617] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.705848] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.785580] env[65107]: WARNING openstack [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.785904] env[65107]: WARNING openstack [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.852769] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103379, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.885899] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updated VIF entry in instance network info cache for port fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1040.886782] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "address": "fa:16:3e:6b:22:2f", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbe601d0-6b", "ovs_interfaceid": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1040.889611] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a69ff93-074e-4c77-a071-97f99cfa1bb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.897814] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc39160-6f82-4749-81f7-a8e2d7f3f440 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.931750] env[65107]: WARNING neutronclient.v2_0.client [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1040.932492] env[65107]: WARNING openstack [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1040.932905] env[65107]: WARNING openstack [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1040.940965] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cf80c6-c187-4fc4-b7ea-c6d67f5cebff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.950660] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe4eb5e-2042-4a74-a2cc-b59a28111a62 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.970355] env[65107]: DEBUG nova.compute.provider_tree [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.182853] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1041.183514] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8bb725b1-80ac-4a38-8647-9e39d4edf8f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.195903] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1041.195903] env[65107]: value = "task-5103380" [ 1041.195903] env[65107]: _type = "Task" [ 1041.195903] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.207231] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103380, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.254768] env[65107]: DEBUG nova.network.neutron [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1041.346252] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103379, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.396015] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1041.396015] env[65107]: DEBUG nova.compute.manager [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1041.396015] env[65107]: DEBUG nova.compute.manager [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing instance network info cache due to event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1041.396015] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.396015] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.396015] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1041.476252] env[65107]: DEBUG nova.scheduler.client.report [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1041.553064] env[65107]: DEBUG nova.compute.manager [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Received event network-changed-96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1041.553269] env[65107]: DEBUG nova.compute.manager [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Refreshing instance network info cache due to event network-changed-96111c75-74f7-4496-9470-0972fef27f6f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1041.553489] env[65107]: DEBUG oslo_concurrency.lockutils [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Acquiring lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.553644] env[65107]: DEBUG oslo_concurrency.lockutils [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Acquired lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.553812] env[65107]: DEBUG nova.network.neutron [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Refreshing network info cache for port 96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1041.707343] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103380, 'name': CloneVM_Task} progress is 93%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.757765] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1041.793881] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c93b8199db546816feb7959d2c1c2e30',container_format='bare',created_at=2025-12-12T13:27:23Z,direct_url=,disk_format='vmdk',id=c38e9e3f-de87-4115-b1da-c44beabd8098,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-417011522-shelved',owner='dced4f9aef2f49cf990203b693533aa1',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2025-12-12T13:27:39Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1041.793881] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1041.794280] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1041.794377] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1041.794618] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1041.794922] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1041.795341] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.795625] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1041.795981] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1041.796364] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1041.796675] env[65107]: DEBUG nova.virt.hardware [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1041.798110] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8d9dca-33c5-4807-9563-ed49490cac37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.816171] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98a6395-1729-4a60-a21f-dc3d6dfc1c84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.842276] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:15:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6bba8b1-799c-4c2b-bd94-8726094948ee', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.856863] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1041.862655] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1041.863377] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c11bffff-6a6c-4864-8c6e-e3c17a111c45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.899892] env[65107]: WARNING neutronclient.v2_0.client [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.901043] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.901689] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.913097] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103379, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.919403] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.919403] env[65107]: value = "task-5103381" [ 1041.919403] env[65107]: _type = "Task" [ 1041.919403] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.927821] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103381, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.984760] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.877s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.989049] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.567s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.989510] env[65107]: DEBUG nova.objects.instance [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1042.022921] env[65107]: INFO nova.scheduler.client.report [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocations for instance b4f7d003-59fb-4ef4-b5f6-234d5154d198 [ 1042.043307] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-57c8daf0-4e71-4f3e-8e92-cde14428fbe3" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.043706] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-57c8daf0-4e71-4f3e-8e92-cde14428fbe3" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.044142] env[65107]: DEBUG nova.objects.instance [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'flavor' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.057793] env[65107]: WARNING neutronclient.v2_0.client [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.058765] env[65107]: WARNING openstack [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.059381] env[65107]: WARNING openstack [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.206513] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.206972] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.226732] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103380, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.254916] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "a5ea4075-616a-49ce-aa93-f1db495a022b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.255347] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.296759] env[65107]: DEBUG nova.compute.manager [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-vif-plugged-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1042.297168] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.297549] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.297852] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.298198] env[65107]: DEBUG nova.compute.manager [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] No waiting events found dispatching network-vif-plugged-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1042.298501] env[65107]: WARNING nova.compute.manager [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received unexpected event network-vif-plugged-b6bba8b1-799c-4c2b-bd94-8726094948ee for instance with vm_state shelved_offloaded and task_state spawning. [ 1042.298834] env[65107]: DEBUG nova.compute.manager [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1042.299250] env[65107]: DEBUG nova.compute.manager [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing instance network info cache due to event network-changed-b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1042.299425] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Acquiring lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.299674] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Acquired lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.299947] env[65107]: DEBUG nova.network.neutron [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Refreshing network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1042.352380] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103379, 'name': CreateVM_Task, 'duration_secs': 1.804434} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.352577] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04a508b4-4430-411c-8573-726065b558a1] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.353095] env[65107]: WARNING neutronclient.v2_0.client [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.353471] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.353622] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.354129] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1042.354478] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ea6c6dd-4718-47fb-8ff9-1346ea15d1fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.361342] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1042.361342] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4e95d-40f3-409d-d114-9ee0260814be" [ 1042.361342] env[65107]: _type = "Task" [ 1042.361342] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.371824] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4e95d-40f3-409d-d114-9ee0260814be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.373330] env[65107]: WARNING neutronclient.v2_0.client [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.374015] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.374526] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.389117] env[65107]: WARNING openstack [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.389636] env[65107]: WARNING openstack [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.428192] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103381, 'name': CreateVM_Task, 'duration_secs': 0.486749} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.428400] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.428935] env[65107]: WARNING neutronclient.v2_0.client [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.429335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.429485] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.429876] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1042.430154] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ff3f5c9-c27d-471b-8ac2-d448a1db57d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.436371] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1042.436371] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a369c-dc2a-26bd-c25d-589058420508" [ 1042.436371] env[65107]: _type = "Task" [ 1042.436371] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.447545] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520a369c-dc2a-26bd-c25d-589058420508, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.536557] env[65107]: DEBUG oslo_concurrency.lockutils [None req-060c7414-f8c7-49c7-b696-09043bd1c093 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "b4f7d003-59fb-4ef4-b5f6-234d5154d198" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.639s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.549010] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.549844] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.550285] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.576442] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updated VIF entry in instance network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1042.576802] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1042.611813] env[65107]: WARNING neutronclient.v2_0.client [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.612760] env[65107]: WARNING openstack [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.613137] env[65107]: WARNING openstack [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.711366] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103380, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.760057] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1042.804877] env[65107]: WARNING neutronclient.v2_0.client [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.805619] env[65107]: WARNING openstack [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.805977] env[65107]: WARNING openstack [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.818827] env[65107]: DEBUG nova.network.neutron [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updated VIF entry in instance network info cache for port 96111c75-74f7-4496-9470-0972fef27f6f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1042.819086] env[65107]: DEBUG nova.network.neutron [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updating instance_info_cache with network_info: [{"id": "96111c75-74f7-4496-9470-0972fef27f6f", "address": "fa:16:3e:ab:6f:69", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96111c75-74", "ovs_interfaceid": "96111c75-74f7-4496-9470-0972fef27f6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1042.860437] env[65107]: DEBUG nova.objects.instance [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'pci_requests' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.874058] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d4e95d-40f3-409d-d114-9ee0260814be, 'name': SearchDatastore_Task, 'duration_secs': 0.014586} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.874850] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1042.876351] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.876675] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.876861] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.877109] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.877815] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efd5d5af-a516-43fe-9e54-ac4293d010ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.888824] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.889034] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1042.890055] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfb14668-22df-4d8e-be8d-fac186d28b1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.896731] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1042.896731] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527a3040-edfd-f30b-ba9c-e998d4971faa" [ 1042.896731] env[65107]: _type = "Task" [ 1042.896731] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.908128] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527a3040-edfd-f30b-ba9c-e998d4971faa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.953928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1042.954170] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Processing image c38e9e3f-de87-4115-b1da-c44beabd8098 {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.954411] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.954559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.954791] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.955172] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fca50a5e-0cff-4011-8fa5-bbdd6e73ad9c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.971108] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.971326] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1042.972314] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b03a0816-10d6-4762-944d-d4b5537aa864 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.979751] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1042.979751] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c0c93-0960-e3ae-b085-a607efe8d5a8" [ 1042.979751] env[65107]: _type = "Task" [ 1042.979751] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.989124] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c0c93-0960-e3ae-b085-a607efe8d5a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.003553] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c49fcc01-1817-4cdb-b367-01bd5afb6fbb tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1043.005348] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 10.950s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1043.027399] env[65107]: WARNING openstack [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.029369] env[65107]: WARNING openstack [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.079871] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.080513] env[65107]: DEBUG nova.compute.manager [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1043.080513] env[65107]: DEBUG nova.compute.manager [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing instance network info cache due to event network-changed-ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1043.080655] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Acquiring lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.080749] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Acquired lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.081129] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Refreshing network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1043.207528] env[65107]: WARNING neutronclient.v2_0.client [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.209051] env[65107]: WARNING openstack [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.209051] env[65107]: WARNING openstack [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.224771] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103380, 'name': CloneVM_Task, 'duration_secs': 1.655195} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.225185] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Created linked-clone VM from snapshot [ 1043.225959] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505269b7-ee9c-43e1-91e2-618f913596bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.235618] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Uploading image d085e822-30a8-4599-bb8b-1435b8d6be97 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1043.251896] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1043.251896] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3f044a35-7321-4f85-97f8-32b80472bab2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.258999] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1043.258999] env[65107]: value = "task-5103382" [ 1043.258999] env[65107]: _type = "Task" [ 1043.258999] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.273599] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103382, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.295895] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1043.326290] env[65107]: DEBUG oslo_concurrency.lockutils [req-b9d05d90-c087-4692-b536-325661eb48ad req-aa67d823-abdd-43b7-9ece-e7b4f1e144b2 service nova] Releasing lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.331585] env[65107]: DEBUG nova.network.neutron [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updated VIF entry in instance network info cache for port b6bba8b1-799c-4c2b-bd94-8726094948ee. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1043.332166] env[65107]: DEBUG nova.network.neutron [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [{"id": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "address": "fa:16:3e:44:15:6b", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6bba8b1-79", "ovs_interfaceid": "b6bba8b1-799c-4c2b-bd94-8726094948ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.366289] env[65107]: DEBUG nova.objects.base [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1043.366713] env[65107]: DEBUG nova.network.neutron [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1043.366913] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.367260] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.368067] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.368457] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.409394] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527a3040-edfd-f30b-ba9c-e998d4971faa, 'name': SearchDatastore_Task, 'duration_secs': 0.011973} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.414106] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37589ad2-8668-4279-9aef-0a9c12e7caa4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.418619] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1043.418619] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52143e13-9ae2-a67f-bf33-9fd1b2cf87c5" [ 1043.418619] env[65107]: _type = "Task" [ 1043.418619] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.427760] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52143e13-9ae2-a67f-bf33-9fd1b2cf87c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.488210] env[65107]: DEBUG nova.policy [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1043.499388] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1043.499661] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Fetch image to [datastore1] OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477/OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1043.499848] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Downloading stream optimized image c38e9e3f-de87-4115-b1da-c44beabd8098 to [datastore1] OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477/OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477.vmdk on the data store datastore1 as vApp {{(pid=65107) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1043.500351] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Downloading image file data c38e9e3f-de87-4115-b1da-c44beabd8098 to the ESX as VM named 'OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477' {{(pid=65107) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1043.511102] env[65107]: INFO nova.compute.claims [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.584709] env[65107]: WARNING neutronclient.v2_0.client [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.585321] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.585726] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.594314] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1043.594314] env[65107]: value = "resgroup-9" [ 1043.594314] env[65107]: _type = "ResourcePool" [ 1043.594314] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1043.594584] env[65107]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-834e9d81-5969-4d61-9a64-919dadc35254 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.616936] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease: (returnval){ [ 1043.616936] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e85ba6-fab8-fb66-6c5d-1cf7b24f376b" [ 1043.616936] env[65107]: _type = "HttpNfcLease" [ 1043.616936] env[65107]: } obtained for vApp import into resource pool (val){ [ 1043.616936] env[65107]: value = "resgroup-9" [ 1043.616936] env[65107]: _type = "ResourcePool" [ 1043.616936] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1043.617456] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the lease: (returnval){ [ 1043.617456] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e85ba6-fab8-fb66-6c5d-1cf7b24f376b" [ 1043.617456] env[65107]: _type = "HttpNfcLease" [ 1043.617456] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1043.624545] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1043.624545] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e85ba6-fab8-fb66-6c5d-1cf7b24f376b" [ 1043.624545] env[65107]: _type = "HttpNfcLease" [ 1043.624545] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1043.769519] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103382, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.787606] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.788014] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.835995] env[65107]: DEBUG oslo_concurrency.lockutils [req-1f6a89af-f8e4-495c-aa3a-60abf2fa068d req-85c7b7f2-cc97-44b7-9cad-073b840ae234 service nova] Releasing lock "refresh_cache-d2e7eee7-c9b2-49aa-92f9-1efd7c901509" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.912762] env[65107]: WARNING neutronclient.v2_0.client [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.913368] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.914080] env[65107]: WARNING openstack [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.935971] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52143e13-9ae2-a67f-bf33-9fd1b2cf87c5, 'name': SearchDatastore_Task, 'duration_secs': 0.06414} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.936131] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.936404] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 04a508b4-4430-411c-8573-726065b558a1/04a508b4-4430-411c-8573-726065b558a1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1043.936686] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8665fa6a-2674-4aeb-9e12-40fb277af5ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.946151] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1043.946151] env[65107]: value = "task-5103384" [ 1043.946151] env[65107]: _type = "Task" [ 1043.946151] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.956678] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.024299] env[65107]: INFO nova.compute.resource_tracker [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating resource usage from migration 387dc95a-dd2c-4f87-8187-faf6d8f7234c [ 1044.029701] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updated VIF entry in instance network info cache for port ed79bc21-5c5f-44e5-ba58-55e43c814088. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1044.030387] env[65107]: DEBUG nova.network.neutron [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [{"id": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "address": "fa:16:3e:6f:25:45", "network": {"id": "9f98695f-bf85-4b65-9a91-a351b6f29d37", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1223383640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfb7fca6907c4c32a1aa6b8023c39afb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped79bc21-5c", "ovs_interfaceid": "ed79bc21-5c5f-44e5-ba58-55e43c814088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.128391] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1044.128391] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e85ba6-fab8-fb66-6c5d-1cf7b24f376b" [ 1044.128391] env[65107]: _type = "HttpNfcLease" [ 1044.128391] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1044.271323] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103382, 'name': Destroy_Task, 'duration_secs': 0.821955} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.274491] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Destroyed the VM [ 1044.274792] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1044.275340] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-efc22f67-d10b-49c1-9132-4211deeb7122 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.286360] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1044.286360] env[65107]: value = "task-5103385" [ 1044.286360] env[65107]: _type = "Task" [ 1044.286360] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.288359] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7191a731-6e94-4510-8bd7-2a4a2435cfbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.305697] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d804e065-ce3e-4a7f-8b66-e1e5abba6fba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.309656] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103385, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.342704] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f224d7db-9a93-4a84-9bcd-f4b5e1b4e12b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.352212] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8329821f-17ac-4130-b77e-e0cac82f3f40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.369501] env[65107]: DEBUG nova.compute.provider_tree [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.458785] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103384, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.533442] env[65107]: DEBUG oslo_concurrency.lockutils [req-fba9863a-a936-489a-860a-2f10f127dec9 req-30362b36-0f06-4434-8109-f36fe55e7802 service nova] Releasing lock "refresh_cache-e12cb11f-7d6f-47d6-a31e-bf2654174d38" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.614312] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "7373fbbe-52c1-4955-8da9-a5563d71391b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.614580] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.630966] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1044.630966] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e85ba6-fab8-fb66-6c5d-1cf7b24f376b" [ 1044.630966] env[65107]: _type = "HttpNfcLease" [ 1044.630966] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1044.631303] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1044.631303] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e85ba6-fab8-fb66-6c5d-1cf7b24f376b" [ 1044.631303] env[65107]: _type = "HttpNfcLease" [ 1044.631303] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1044.632023] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b4628a-07b7-4333-b65f-345e8bb7e15e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.641178] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528c5d15-454b-28cd-7629-0963ba7499ce/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1044.641373] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528c5d15-454b-28cd-7629-0963ba7499ce/disk-0.vmdk. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1044.715706] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-249915d8-7bf6-479b-8029-b0189cc101d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.798374] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103385, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.873062] env[65107]: DEBUG nova.scheduler.client.report [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1044.969721] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103384, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566704} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.969721] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 04a508b4-4430-411c-8573-726065b558a1/04a508b4-4430-411c-8573-726065b558a1.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.969721] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.969721] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d79ac664-d1e8-4698-aa1d-5a5c8b30c30b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.980086] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1044.980086] env[65107]: value = "task-5103386" [ 1044.980086] env[65107]: _type = "Task" [ 1044.980086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.998787] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.034697] env[65107]: DEBUG nova.compute.manager [req-fd154c8b-8a46-4e1e-94d9-6ee9fb486be1 req-fb9f7bc1-89d5-45aa-8606-594e112a89db service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-vif-plugged-57c8daf0-4e71-4f3e-8e92-cde14428fbe3 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1045.035182] env[65107]: DEBUG oslo_concurrency.lockutils [req-fd154c8b-8a46-4e1e-94d9-6ee9fb486be1 req-fb9f7bc1-89d5-45aa-8606-594e112a89db service nova] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.035182] env[65107]: DEBUG oslo_concurrency.lockutils [req-fd154c8b-8a46-4e1e-94d9-6ee9fb486be1 req-fb9f7bc1-89d5-45aa-8606-594e112a89db service nova] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.035504] env[65107]: DEBUG oslo_concurrency.lockutils [req-fd154c8b-8a46-4e1e-94d9-6ee9fb486be1 req-fb9f7bc1-89d5-45aa-8606-594e112a89db service nova] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.035702] env[65107]: DEBUG nova.compute.manager [req-fd154c8b-8a46-4e1e-94d9-6ee9fb486be1 req-fb9f7bc1-89d5-45aa-8606-594e112a89db service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] No waiting events found dispatching network-vif-plugged-57c8daf0-4e71-4f3e-8e92-cde14428fbe3 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1045.036039] env[65107]: WARNING nova.compute.manager [req-fd154c8b-8a46-4e1e-94d9-6ee9fb486be1 req-fb9f7bc1-89d5-45aa-8606-594e112a89db service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received unexpected event network-vif-plugged-57c8daf0-4e71-4f3e-8e92-cde14428fbe3 for instance with vm_state active and task_state None. [ 1045.113222] env[65107]: DEBUG nova.network.neutron [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Successfully updated port: 57c8daf0-4e71-4f3e-8e92-cde14428fbe3 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1045.117485] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1045.302377] env[65107]: DEBUG oslo_vmware.api [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103385, 'name': RemoveSnapshot_Task, 'duration_secs': 0.89651} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.304150] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1045.378287] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.373s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.378549] env[65107]: INFO nova.compute.manager [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Migrating [ 1045.387703] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.071s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.387965] env[65107]: DEBUG nova.objects.instance [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lazy-loading 'resources' on Instance uuid 888d60d4-bc9c-48af-97f9-434883884e8a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.493571] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077167} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.493888] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.494793] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30205bac-bd75-4883-b359-d743d699a240 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.522561] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 04a508b4-4430-411c-8573-726065b558a1/04a508b4-4430-411c-8573-726065b558a1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.527308] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96f76598-2086-449b-8c77-00047772a34e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.557070] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1045.557070] env[65107]: value = "task-5103387" [ 1045.557070] env[65107]: _type = "Task" [ 1045.557070] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.568853] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103387, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.615618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.616050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.616117] env[65107]: DEBUG nova.network.neutron [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1045.639439] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.810061] env[65107]: WARNING nova.compute.manager [None req-043a2616-6af1-4a13-a831-42e4eea63e8e tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Image not found during snapshot: nova.exception.ImageNotFound: Image d085e822-30a8-4599-bb8b-1435b8d6be97 could not be found. [ 1045.909559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.909559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.909780] env[65107]: DEBUG nova.network.neutron [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1045.927562] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1045.927797] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528c5d15-454b-28cd-7629-0963ba7499ce/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1045.929190] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba29b32-75e4-4fb4-8758-991ba7f7c0d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.939855] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528c5d15-454b-28cd-7629-0963ba7499ce/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1045.940074] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528c5d15-454b-28cd-7629-0963ba7499ce/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1045.940396] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9e96444f-f2fa-4f9f-84a8-c58048738d46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.035773] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "8dbed811-3d4e-4cae-9981-0334801013aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.036150] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.073960] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103387, 'name': ReconfigVM_Task, 'duration_secs': 0.420131} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.074204] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 04a508b4-4430-411c-8573-726065b558a1/04a508b4-4430-411c-8573-726065b558a1.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.077591] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dff6b0fb-e306-45e7-8f61-c21ee2977b07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.087435] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1046.087435] env[65107]: value = "task-5103388" [ 1046.087435] env[65107]: _type = "Task" [ 1046.087435] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.105230] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103388, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.119187] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.119572] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.158548] env[65107]: DEBUG oslo_vmware.rw_handles [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528c5d15-454b-28cd-7629-0963ba7499ce/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1046.159333] env[65107]: INFO nova.virt.vmwareapi.images [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Downloaded image file data c38e9e3f-de87-4115-b1da-c44beabd8098 [ 1046.159635] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e391083-a282-4d86-aa3e-bfb652bdbbbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.181723] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f299ea7-f425-4e82-b602-84dd91d7a2b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.199115] env[65107]: WARNING nova.network.neutron [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] 40a76aa5-ac46-4067-a98d-cd60410bf479 already exists in list: networks containing: ['40a76aa5-ac46-4067-a98d-cd60410bf479']. ignoring it [ 1046.199581] env[65107]: WARNING nova.network.neutron [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] 40a76aa5-ac46-4067-a98d-cd60410bf479 already exists in list: networks containing: ['40a76aa5-ac46-4067-a98d-cd60410bf479']. ignoring it [ 1046.211340] env[65107]: INFO nova.virt.vmwareapi.images [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] The imported VM was unregistered [ 1046.214938] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1046.215285] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.216277] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bab1717f-793a-4a3e-8d85-5cd660eba082 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.231289] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created directory with path [datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.231584] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477/OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477.vmdk to [datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk. {{(pid=65107) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1046.231871] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-41097e48-57ac-4cd2-8c3c-97b7634b4b29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.242263] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1046.242263] env[65107]: value = "task-5103390" [ 1046.242263] env[65107]: _type = "Task" [ 1046.242263] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.250788] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.266476] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae566785-a589-43cf-8c8b-44617a440e59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.274320] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b63919e-5690-475c-ae6c-f30e2f019e02 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.307770] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.308197] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.315782] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e789845f-121c-4557-bb91-40dbf7e136d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.326835] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60f145b-63fa-4faa-910f-d14d59b96d8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.343378] env[65107]: DEBUG nova.compute.provider_tree [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.397580] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.398274] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.398617] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.414041] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.414709] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.415155] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.542577] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1046.571870] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.572556] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.582099] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.582350] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.590248] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.590248] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.617165] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103388, 'name': Rename_Task, 'duration_secs': 0.178134} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.617165] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.617422] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3112e74a-b87c-40bb-9207-3a681d0393eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.630447] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1046.630447] env[65107]: value = "task-5103391" [ 1046.630447] env[65107]: _type = "Task" [ 1046.630447] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.643941] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.650556] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.651352] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.651821] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.663158] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.663756] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.664116] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.758845] env[65107]: DEBUG nova.network.neutron [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [{"id": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "address": "fa:16:3e:c6:73:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd88e66c9-ba", "ovs_interfaceid": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1046.764280] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.777227] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.777763] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.847309] env[65107]: DEBUG nova.scheduler.client.report [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1046.851964] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.852642] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.852985] env[65107]: WARNING openstack [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.946185] env[65107]: DEBUG nova.network.neutron [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "address": "fa:16:3e:6b:22:2f", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbe601d0-6b", "ovs_interfaceid": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "57c8daf0-4e71-4f3e-8e92-cde14428fbe3", "address": "fa:16:3e:21:8c:18", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57c8daf0-4e", "ovs_interfaceid": "57c8daf0-4e71-4f3e-8e92-cde14428fbe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1047.068418] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.086108] env[65107]: DEBUG nova.compute.utils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1047.146701] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.255903] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.265731] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.364299] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.366905] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.015s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.367272] env[65107]: DEBUG nova.objects.instance [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lazy-loading 'resources' on Instance uuid a3162257-ac6b-4468-a573-3611bba69c46 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.389009] env[65107]: INFO nova.scheduler.client.report [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Deleted allocations for instance 888d60d4-bc9c-48af-97f9-434883884e8a [ 1047.449160] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.449832] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.449989] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.451178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e29de9-6632-404c-b974-dde97b2eb32e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.474336] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1047.474585] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1047.474718] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1047.474910] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1047.475097] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1047.475270] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1047.475476] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.475629] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1047.475786] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1047.475940] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1047.476127] env[65107]: DEBUG nova.virt.hardware [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1047.482590] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfiguring VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1047.482995] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b4ec66a-5dd7-4f85-8bf5-07e927ba10e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.503627] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1047.503627] env[65107]: value = "task-5103392" [ 1047.503627] env[65107]: _type = "Task" [ 1047.503627] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.515243] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103392, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.589914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.644940] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.757432] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.897139] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6f415f8-9dca-4374-911c-53af96ae967e tempest-ServersAdminNegativeTestJSON-1381091624 tempest-ServersAdminNegativeTestJSON-1381091624-project-member] Lock "888d60d4-bc9c-48af-97f9-434883884e8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.014s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.020746] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103392, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.117481] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e863994-c663-4aef-9b7f-7c40265f2f4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.131596] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139f7e78-aa3c-42cb-b8a4-a87cd2421741 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.154067] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.186225] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9f8969-9020-45b3-bcd5-82a817c6a313 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.201345] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f56f7a0-64a7-4ec8-a215-8116da6c023b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.219766] env[65107]: DEBUG nova.compute.provider_tree [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.258200] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.519290] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103392, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.649449] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.686749] env[65107]: DEBUG nova.compute.manager [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-changed-57c8daf0-4e71-4f3e-8e92-cde14428fbe3 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1048.686891] env[65107]: DEBUG nova.compute.manager [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing instance network info cache due to event network-changed-57c8daf0-4e71-4f3e-8e92-cde14428fbe3. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1048.687111] env[65107]: DEBUG oslo_concurrency.lockutils [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.687251] env[65107]: DEBUG oslo_concurrency.lockutils [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.687430] env[65107]: DEBUG nova.network.neutron [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Refreshing network info cache for port 57c8daf0-4e71-4f3e-8e92-cde14428fbe3 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1048.693054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.693354] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.693598] env[65107]: INFO nova.compute.manager [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Attaching volume be6f9921-262e-4dcb-bc60-481f0ad36a02 to /dev/sdb [ 1048.723376] env[65107]: DEBUG nova.scheduler.client.report [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1048.730635] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8477f21-a2bd-4ce3-89c5-e8eedaea6f41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.739454] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6944857d-9315-45f0-8ff6-35722dd07488 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.757521] env[65107]: DEBUG nova.virt.block_device [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating existing volume attachment record: b61a26cf-19c9-4960-876d-015d83015ced {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1048.765313] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.765579] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.765792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.765969] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.766148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.767812] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.768411] env[65107]: INFO nova.compute.manager [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Terminating instance [ 1048.783651] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a6b778-3110-4286-b14f-d93bb13db523 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.803809] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.015352] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103392, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.148442] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.191363] env[65107]: WARNING neutronclient.v2_0.client [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.191363] env[65107]: WARNING openstack [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.191907] env[65107]: WARNING openstack [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.236770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.240042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.943s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.241172] env[65107]: INFO nova.compute.claims [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.262774] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103390, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.608275} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.263125] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477/OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477.vmdk to [datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk. [ 1049.263399] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Cleaning up location [datastore1] OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477 {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1049.263723] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_3dbdb774-6ec5-485b-a22b-96031bf98477 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.263828] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-045d5edc-c3bb-4e3f-8047-d3eedd73315c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.266848] env[65107]: INFO nova.scheduler.client.report [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Deleted allocations for instance a3162257-ac6b-4468-a573-3611bba69c46 [ 1049.272529] env[65107]: DEBUG nova.compute.manager [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1049.272959] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.276402] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a43e6a0-cb9b-4574-b38b-b3694dc454c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.278611] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1049.278611] env[65107]: value = "task-5103396" [ 1049.278611] env[65107]: _type = "Task" [ 1049.278611] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.284744] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.285935] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fcad689-c0c5-49e2-9278-146bb5857970 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.290959] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.293150] env[65107]: DEBUG oslo_vmware.api [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1049.293150] env[65107]: value = "task-5103397" [ 1049.293150] env[65107]: _type = "Task" [ 1049.293150] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.302023] env[65107]: DEBUG oslo_vmware.api [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103397, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.312497] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.312831] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7852e1a-09f3-40c7-976c-71618e019061 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.320057] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1049.320057] env[65107]: value = "task-5103398" [ 1049.320057] env[65107]: _type = "Task" [ 1049.320057] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.330827] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103398, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.332790] env[65107]: WARNING openstack [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.333386] env[65107]: WARNING openstack [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.419137] env[65107]: WARNING neutronclient.v2_0.client [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.419896] env[65107]: WARNING openstack [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.420407] env[65107]: WARNING openstack [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.522926] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103392, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.525649] env[65107]: DEBUG nova.network.neutron [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updated VIF entry in instance network info cache for port 57c8daf0-4e71-4f3e-8e92-cde14428fbe3. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1049.526246] env[65107]: DEBUG nova.network.neutron [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "address": "fa:16:3e:6b:22:2f", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbe601d0-6b", "ovs_interfaceid": "fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "57c8daf0-4e71-4f3e-8e92-cde14428fbe3", "address": "fa:16:3e:21:8c:18", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57c8daf0-4e", "ovs_interfaceid": "57c8daf0-4e71-4f3e-8e92-cde14428fbe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.650928] env[65107]: DEBUG oslo_vmware.api [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103391, 'name': PowerOnVM_Task, 'duration_secs': 2.846457} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.651243] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1049.651454] env[65107]: INFO nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Took 12.82 seconds to spawn the instance on the hypervisor. [ 1049.651634] env[65107]: DEBUG nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1049.652459] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d8745b-4cec-4350-ac97-f60eb293eb35 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.775450] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c4c23692-2d8f-4dd4-a23a-5293d20c39d6 tempest-ServersListShow2100Test-1231567060 tempest-ServersListShow2100Test-1231567060-project-member] Lock "a3162257-ac6b-4468-a573-3611bba69c46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.907s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.790194] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.077393} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.790456] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.790621] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.790875] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk to [datastore1] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.791166] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e17cf5e-b28a-4f9f-841e-2b404f991f1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.804023] env[65107]: DEBUG oslo_vmware.api [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103397, 'name': PowerOffVM_Task, 'duration_secs': 0.397254} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.805598] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.805772] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.806422] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1049.806422] env[65107]: value = "task-5103399" [ 1049.806422] env[65107]: _type = "Task" [ 1049.806422] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.806633] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85711e34-fc68-43b3-9c25-bbe6a1d046cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.817881] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103399, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.831929] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103398, 'name': PowerOffVM_Task, 'duration_secs': 0.226647} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.832232] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.832412] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.897521] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.897766] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.898159] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleting the datastore file [datastore1] d67468c6-9aaf-49c4-afe4-5d9856c4af7a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.898903] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7827c232-bc03-4e7d-b6cc-735a473c4ca4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.907227] env[65107]: DEBUG oslo_vmware.api [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for the task: (returnval){ [ 1049.907227] env[65107]: value = "task-5103401" [ 1049.907227] env[65107]: _type = "Task" [ 1049.907227] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.916974] env[65107]: DEBUG oslo_vmware.api [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.018779] env[65107]: DEBUG oslo_vmware.api [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103392, 'name': ReconfigVM_Task, 'duration_secs': 2.101098} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.019367] env[65107]: WARNING neutronclient.v2_0.client [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.019656] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.019954] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfigured VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1050.029653] env[65107]: DEBUG oslo_concurrency.lockutils [req-6afd5a4b-9cf7-4926-849e-d5734da6cee6 req-28162729-9b8f-4114-a2b0-02fbdbb08840 service nova] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.175452] env[65107]: INFO nova.compute.manager [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Took 41.28 seconds to build instance. [ 1050.327132] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103399, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.340362] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1050.340701] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1050.340867] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1050.341141] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1050.341372] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1050.341643] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1050.342026] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.342274] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1050.342600] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1050.342886] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1050.343189] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1050.352726] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-668f2508-0499-4e68-a101-d1464d8e2f6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.373461] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1050.373461] env[65107]: value = "task-5103402" [ 1050.373461] env[65107]: _type = "Task" [ 1050.373461] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.390262] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103402, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.423550] env[65107]: DEBUG oslo_vmware.api [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Task: {'id': task-5103401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.513758} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.423871] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.424166] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.424350] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.424480] env[65107]: INFO nova.compute.manager [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1050.424853] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1050.425203] env[65107]: DEBUG nova.compute.manager [-] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1050.425279] env[65107]: DEBUG nova.network.neutron [-] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1050.425597] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.426213] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.426479] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.507735] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.525181] env[65107]: DEBUG oslo_concurrency.lockutils [None req-86f73216-c31e-48d1-80e9-ee1c1aea39fe tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-57c8daf0-4e71-4f3e-8e92-cde14428fbe3" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.481s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.554710] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f195e0e-e4eb-473b-84f3-2fe87b651155 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.569091] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946fa5e2-ef0d-4f32-9766-7f93c3a395ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.606596] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff494b74-e401-4f67-9a29-1584d4d823c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.617313] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8863fc1d-c988-4643-87ea-adeb19124194 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.636365] env[65107]: DEBUG nova.compute.provider_tree [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1050.677264] env[65107]: DEBUG oslo_concurrency.lockutils [None req-76c2f980-c06e-4c7f-8bc1-688a5fef9058 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.796s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.827025] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103399, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.887160] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103402, 'name': ReconfigVM_Task, 'duration_secs': 0.34039} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.887588] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1051.007309] env[65107]: DEBUG nova.compute.manager [req-4108d8a0-8ccb-4b3f-81f2-3e37ccd7776e req-ab833e31-4fed-4850-aec4-fc443fddb9e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Received event network-vif-deleted-4fe4b3ad-385c-4532-8edc-066c218ced49 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1051.007849] env[65107]: INFO nova.compute.manager [req-4108d8a0-8ccb-4b3f-81f2-3e37ccd7776e req-ab833e31-4fed-4850-aec4-fc443fddb9e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Neutron deleted interface 4fe4b3ad-385c-4532-8edc-066c218ced49; detaching it from the instance and deleting it from the info cache [ 1051.007849] env[65107]: DEBUG nova.network.neutron [req-4108d8a0-8ccb-4b3f-81f2-3e37ccd7776e req-ab833e31-4fed-4850-aec4-fc443fddb9e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1051.166369] env[65107]: ERROR nova.scheduler.client.report [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [req-6d0d87cd-8ff6-464b-b6e9-0551ae1a63d7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8adccf8d-6d73-43c0-ad15-ee89967ab8cd. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6d0d87cd-8ff6-464b-b6e9-0551ae1a63d7"}]} [ 1051.188985] env[65107]: DEBUG nova.scheduler.client.report [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1051.210508] env[65107]: DEBUG nova.scheduler.client.report [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1051.210740] env[65107]: DEBUG nova.compute.provider_tree [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.226285] env[65107]: DEBUG nova.scheduler.client.report [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1051.246027] env[65107]: DEBUG nova.scheduler.client.report [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1051.285440] env[65107]: DEBUG nova.network.neutron [-] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1051.326767] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103399, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.397225] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1051.397515] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1051.397671] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1051.397889] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1051.397990] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1051.399301] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1051.399540] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.400217] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1051.400666] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1051.400666] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1051.400806] env[65107]: DEBUG nova.virt.hardware [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1051.409345] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1051.412689] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f19cd81c-7fd0-4ff0-bd27-ce0b87b98176 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.435255] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1051.435255] env[65107]: value = "task-5103404" [ 1051.435255] env[65107]: _type = "Task" [ 1051.435255] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.448527] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103404, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.511581] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76ccfa49-09b1-40c9-8908-605acf5d30c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.525234] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e71bc9-c7f5-4479-9541-e0caa67a7198 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.571798] env[65107]: DEBUG nova.compute.manager [req-4108d8a0-8ccb-4b3f-81f2-3e37ccd7776e req-ab833e31-4fed-4850-aec4-fc443fddb9e2 service nova] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Detach interface failed, port_id=4fe4b3ad-385c-4532-8edc-066c218ced49, reason: Instance d67468c6-9aaf-49c4-afe4-5d9856c4af7a could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1051.574408] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab604d4-4827-44fd-b832-4aa4b7e6dc93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.585159] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef7cfd8-0e36-4e47-8471-0c64cc63ca47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.624190] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79dd562-9614-44ec-9798-253a2d0112e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.634594] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90df513-9f8a-4003-8f7c-532566b87340 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.544467] env[65107]: INFO nova.compute.manager [-] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Took 2.12 seconds to deallocate network for instance. [ 1052.548054] env[65107]: DEBUG nova.compute.manager [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Received event network-changed-96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1052.548257] env[65107]: DEBUG nova.compute.manager [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Refreshing instance network info cache due to event network-changed-96111c75-74f7-4496-9470-0972fef27f6f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1052.548790] env[65107]: DEBUG oslo_concurrency.lockutils [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Acquiring lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.548790] env[65107]: DEBUG oslo_concurrency.lockutils [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Acquired lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1052.548790] env[65107]: DEBUG nova.network.neutron [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Refreshing network info cache for port 96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1052.569545] env[65107]: DEBUG nova.compute.provider_tree [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.577597] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103404, 'name': ReconfigVM_Task, 'duration_secs': 0.421223} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.578212] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103399, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.717729} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.578367] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1052.579211] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c38e9e3f-de87-4115-b1da-c44beabd8098/c38e9e3f-de87-4115-b1da-c44beabd8098.vmdk to [datastore1] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.579459] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ada097-de21-4795-af46-55b9f63c70e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.582524] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e346e0-5346-4d2d-bf72-f877ec86d424 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.608140] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65/00574b77-dad6-4f0a-bbcc-20a2b4a5df65.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.618462] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ce1f1c6-a872-4bbe-9e53-3eb7d570299d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.639403] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.640537] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef2c7ab0-3f9f-4885-91db-4b907705be6a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.661102] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1052.661102] env[65107]: value = "task-5103405" [ 1052.661102] env[65107]: _type = "Task" [ 1052.661102] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.662550] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1052.662550] env[65107]: value = "task-5103406" [ 1052.662550] env[65107]: _type = "Task" [ 1052.662550] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.675515] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103405, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.678886] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103406, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.060577] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.060979] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1053.062783] env[65107]: WARNING neutronclient.v2_0.client [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.063475] env[65107]: WARNING openstack [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.064362] env[65107]: WARNING openstack [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.072894] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.073680] env[65107]: DEBUG nova.scheduler.client.report [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1053.176984] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103405, 'name': ReconfigVM_Task, 'duration_secs': 0.282987} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.179876] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65/00574b77-dad6-4f0a-bbcc-20a2b4a5df65.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.180155] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.183462] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103406, 'name': ReconfigVM_Task, 'duration_secs': 0.297447} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.183913] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfigured VM instance instance-00000048 to attach disk [datastore1] d2e7eee7-c9b2-49aa-92f9-1efd7c901509/d2e7eee7-c9b2-49aa-92f9-1efd7c901509.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.185418] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'boot_index': 0, 'encrypted': False, 'device_name': '/dev/sda', 'size': 0, 'encryption_secret_uuid': None, 'encryption_options': None, 'encryption_format': None, 'guest_format': None, 'device_type': 'disk', 'image_id': 'e439aeac-2298-462c-be63-8218195135cf'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992829', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'name': 'volume-b9b6229b-f202-42ba-8d20-88f1554891f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'd2e7eee7-c9b2-49aa-92f9-1efd7c901509', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'serial': 'b9b6229b-f202-42ba-8d20-88f1554891f7'}, 'disk_bus': None, 'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'attachment_id': 'c9b36cfe-71bf-431b-94dd-0dccd26fcfcf', 'mount_device': '/dev/sdb', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=65107) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1053.185621] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1053.185808] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992829', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'name': 'volume-b9b6229b-f202-42ba-8d20-88f1554891f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'd2e7eee7-c9b2-49aa-92f9-1efd7c901509', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'serial': 'b9b6229b-f202-42ba-8d20-88f1554891f7'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1053.186611] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03591ae4-58a9-46ea-8708-c65a97c20663 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.208916] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee39089c-8728-422c-946c-854897cb6dee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.236242] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] volume-b9b6229b-f202-42ba-8d20-88f1554891f7/volume-b9b6229b-f202-42ba-8d20-88f1554891f7.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.236912] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-702b2e8a-a098-4a20-8cfc-fea7c19c9815 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.255713] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1053.255713] env[65107]: value = "task-5103407" [ 1053.255713] env[65107]: _type = "Task" [ 1053.255713] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.261258] env[65107]: WARNING openstack [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.261661] env[65107]: WARNING openstack [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.272370] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.305364] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1053.305615] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992848', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'name': 'volume-be6f9921-262e-4dcb-bc60-481f0ad36a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8cddc4d1-c94c-4f2b-9c4b-3125e4638050', 'attached_at': '', 'detached_at': '', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'serial': 'be6f9921-262e-4dcb-bc60-481f0ad36a02'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1053.306656] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a6cbe2-e259-4dd0-a6ec-57d30164af8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.323703] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f5554f-1815-454c-bf1a-3bc6b8998686 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.351532] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] volume-be6f9921-262e-4dcb-bc60-481f0ad36a02/volume-be6f9921-262e-4dcb-bc60-481f0ad36a02.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.352867] env[65107]: WARNING neutronclient.v2_0.client [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.353477] env[65107]: WARNING openstack [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.353810] env[65107]: WARNING openstack [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.361060] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57a1dae2-a4ed-413a-8d0f-967da5c9ff83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.380236] env[65107]: DEBUG oslo_vmware.api [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1053.380236] env[65107]: value = "task-5103408" [ 1053.380236] env[65107]: _type = "Task" [ 1053.380236] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.389765] env[65107]: DEBUG oslo_vmware.api [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103408, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.452226] env[65107]: DEBUG nova.network.neutron [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updated VIF entry in instance network info cache for port 96111c75-74f7-4496-9470-0972fef27f6f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1053.452674] env[65107]: DEBUG nova.network.neutron [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updating instance_info_cache with network_info: [{"id": "96111c75-74f7-4496-9470-0972fef27f6f", "address": "fa:16:3e:ab:6f:69", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96111c75-74", "ovs_interfaceid": "96111c75-74f7-4496-9470-0972fef27f6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1053.572037] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.573331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.573331] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cc64fe-8a47-4027-8974-3d5396fbb553 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.579324] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.340s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1053.579933] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1053.582807] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.944s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1053.584227] env[65107]: INFO nova.compute.claims [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.605866] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f17f6b-07d9-423a-acc4-1e0d26773732 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.635200] env[65107]: WARNING neutronclient.v2_0.client [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.640592] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfiguring VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1053.643347] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71991ef2-3719-4119-a906-dd5b936d1344 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.662684] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1053.662684] env[65107]: value = "task-5103409" [ 1053.662684] env[65107]: _type = "Task" [ 1053.662684] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.672269] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.687955] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c10d91-bd7b-4f89-aa88-c1a7ba88a73c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.711618] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ab6bcf-3576-458b-9fc7-47b0e8e4995b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.731305] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.767688] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103407, 'name': ReconfigVM_Task, 'duration_secs': 0.320068} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.768061] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfigured VM instance instance-00000048 to attach disk [datastore1] volume-b9b6229b-f202-42ba-8d20-88f1554891f7/volume-b9b6229b-f202-42ba-8d20-88f1554891f7.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.772852] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ab56970-da9a-4be6-96e2-de16c34db2b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.789424] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1053.789424] env[65107]: value = "task-5103410" [ 1053.789424] env[65107]: _type = "Task" [ 1053.789424] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.801791] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103410, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.890344] env[65107]: DEBUG oslo_vmware.api [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103408, 'name': ReconfigVM_Task, 'duration_secs': 0.397337} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.890637] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to attach disk [datastore1] volume-be6f9921-262e-4dcb-bc60-481f0ad36a02/volume-be6f9921-262e-4dcb-bc60-481f0ad36a02.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.895361] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aee6e2fe-02af-4b27-a5a4-589ce488aa94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.911668] env[65107]: DEBUG oslo_vmware.api [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1053.911668] env[65107]: value = "task-5103411" [ 1053.911668] env[65107]: _type = "Task" [ 1053.911668] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.922309] env[65107]: DEBUG oslo_vmware.api [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103411, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.956283] env[65107]: DEBUG oslo_concurrency.lockutils [req-31e66327-f350-42b8-bdfb-9ab2ecd1b879 req-7a8983d1-d223-4f35-a82c-b09ed976b10d service nova] Releasing lock "refresh_cache-04a508b4-4430-411c-8573-726065b558a1" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1054.093255] env[65107]: DEBUG nova.compute.utils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1054.094698] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1054.094895] env[65107]: DEBUG nova.network.neutron [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1054.095690] env[65107]: WARNING neutronclient.v2_0.client [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.096041] env[65107]: WARNING neutronclient.v2_0.client [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.096688] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.097144] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.154998] env[65107]: DEBUG nova.policy [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1054.174229] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.238353] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.239381] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.300987] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103410, 'name': ReconfigVM_Task, 'duration_secs': 0.156306} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.301322] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992829', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'name': 'volume-b9b6229b-f202-42ba-8d20-88f1554891f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'd2e7eee7-c9b2-49aa-92f9-1efd7c901509', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'serial': 'b9b6229b-f202-42ba-8d20-88f1554891f7'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1054.301903] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caf66910-51c3-4022-8c1a-65bd3d13a673 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.310645] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1054.310645] env[65107]: value = "task-5103412" [ 1054.310645] env[65107]: _type = "Task" [ 1054.310645] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.319479] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103412, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.422315] env[65107]: DEBUG oslo_vmware.api [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103411, 'name': ReconfigVM_Task, 'duration_secs': 0.165994} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.422896] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992848', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'name': 'volume-be6f9921-262e-4dcb-bc60-481f0ad36a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8cddc4d1-c94c-4f2b-9c4b-3125e4638050', 'attached_at': '', 'detached_at': '', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'serial': 'be6f9921-262e-4dcb-bc60-481f0ad36a02'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1054.455635] env[65107]: DEBUG nova.network.neutron [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Successfully created port: 2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1054.514718] env[65107]: DEBUG nova.network.neutron [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Port d88e66c9-ba7c-452d-8e44-8c9603f2b258 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1054.605450] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1054.675414] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.827848] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103412, 'name': Rename_Task, 'duration_secs': 0.26717} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.828608] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.829515] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f090ac21-0c6a-4469-9757-e1a7e02c1849 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.840020] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1054.840020] env[65107]: value = "task-5103413" [ 1054.840020] env[65107]: _type = "Task" [ 1054.840020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.845995] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103413, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.916450] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e78a02-c7d8-405a-bd5b-de1c342cf264 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.929894] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1db2434-8cf1-4ef6-a850-82a2c5c0d9d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.966555] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8d120b-93d1-4879-bc69-d241ff55f5db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.979398] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acb3762-8d43-4bc6-b6ae-840a9514b42f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.002355] env[65107]: DEBUG nova.compute.provider_tree [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.176538] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.349323] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103413, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.473276] env[65107]: DEBUG nova.objects.instance [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'flavor' on Instance uuid 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.505064] env[65107]: DEBUG nova.scheduler.client.report [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1055.540050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.540436] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.540680] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.615623] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1055.650369] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1055.650640] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1055.650795] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1055.650972] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1055.651243] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1055.651345] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1055.651548] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.651745] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1055.651863] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1055.652031] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1055.652203] env[65107]: DEBUG nova.virt.hardware [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1055.653105] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3d6f8f-f05e-4a22-9dd1-3cf2c6957327 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.662055] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ad020b-108e-4001-b5a9-c2372219132c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.676543] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.849999] env[65107]: DEBUG oslo_vmware.api [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103413, 'name': PowerOnVM_Task, 'duration_secs': 0.596674} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.850346] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.924779] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "f6536493-fb80-4fb1-9ee1-e530734e9adf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.925466] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.948339] env[65107]: DEBUG nova.compute.manager [req-06b9c347-49e2-4a31-b328-62861f8f30b0 req-0edf24d9-0370-43d0-8c3d-d1dd0553f582 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Received event network-vif-plugged-2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1055.948339] env[65107]: DEBUG oslo_concurrency.lockutils [req-06b9c347-49e2-4a31-b328-62861f8f30b0 req-0edf24d9-0370-43d0-8c3d-d1dd0553f582 service nova] Acquiring lock "a5ea4075-616a-49ce-aa93-f1db495a022b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.948425] env[65107]: DEBUG oslo_concurrency.lockutils [req-06b9c347-49e2-4a31-b328-62861f8f30b0 req-0edf24d9-0370-43d0-8c3d-d1dd0553f582 service nova] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.948540] env[65107]: DEBUG oslo_concurrency.lockutils [req-06b9c347-49e2-4a31-b328-62861f8f30b0 req-0edf24d9-0370-43d0-8c3d-d1dd0553f582 service nova] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.948711] env[65107]: DEBUG nova.compute.manager [req-06b9c347-49e2-4a31-b328-62861f8f30b0 req-0edf24d9-0370-43d0-8c3d-d1dd0553f582 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] No waiting events found dispatching network-vif-plugged-2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1055.948881] env[65107]: WARNING nova.compute.manager [req-06b9c347-49e2-4a31-b328-62861f8f30b0 req-0edf24d9-0370-43d0-8c3d-d1dd0553f582 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Received unexpected event network-vif-plugged-2cf26308-e979-456c-92a8-f2bdf0d6948f for instance with vm_state building and task_state spawning. [ 1055.978145] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d29d27b7-0b12-4b78-aed1-052cb10a93c0 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.285s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.011890] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.012448] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1056.016765] env[65107]: DEBUG nova.compute.manager [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1056.018050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.949s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.018725] env[65107]: INFO nova.compute.claims [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.021944] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9b141b-0a0a-47a8-8692-d3dea8911af3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.074235] env[65107]: DEBUG nova.network.neutron [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Successfully updated port: 2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1056.176405] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.427615] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1056.530050] env[65107]: DEBUG nova.compute.utils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1056.534168] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1056.534168] env[65107]: DEBUG nova.network.neutron [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1056.534168] env[65107]: WARNING neutronclient.v2_0.client [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.534389] env[65107]: WARNING neutronclient.v2_0.client [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.535106] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.535459] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.553430] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.556603] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5fc157f3-815d-47fa-97bb-c51294c0a649 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 48.202s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.579073] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-a5ea4075-616a-49ce-aa93-f1db495a022b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.579226] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-a5ea4075-616a-49ce-aa93-f1db495a022b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.579528] env[65107]: DEBUG nova.network.neutron [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1056.588383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.589307] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.589307] env[65107]: DEBUG nova.network.neutron [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1056.593768] env[65107]: DEBUG nova.policy [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ccab6477dd44878f78a9d6428d3cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4778fe9152224fd29d1f6220a19b5a36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1056.681477] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.844906] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbd0033-f358-4e77-9b5a-4bf44730f8bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.855211] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae61fa22-2515-487e-9a06-f9e4c186d875 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.893755] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5417e130-6382-4c65-b2b8-5e1d07d0a944 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.903176] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16df3add-6ae5-4647-9d1e-c7923e5253ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.919260] env[65107]: DEBUG nova.compute.provider_tree [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.921896] env[65107]: DEBUG nova.network.neutron [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Successfully created port: 9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1056.933215] env[65107]: DEBUG nova.compute.manager [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1056.955864] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.035045] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1057.082802] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.083468] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.097492] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.098261] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.098641] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.123139] env[65107]: DEBUG nova.network.neutron [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1057.178831] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.209439] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.209851] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.366650] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.367199] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.396911] env[65107]: WARNING neutronclient.v2_0.client [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.397639] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.397979] env[65107]: WARNING openstack [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.429026] env[65107]: DEBUG nova.scheduler.client.report [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1057.468076] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.504849] env[65107]: WARNING neutronclient.v2_0.client [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.505563] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.505909] env[65107]: WARNING openstack [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.544050] env[65107]: DEBUG nova.network.neutron [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Updating instance_info_cache with network_info: [{"id": "2cf26308-e979-456c-92a8-f2bdf0d6948f", "address": "fa:16:3e:ae:d5:48", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf26308-e9", "ovs_interfaceid": "2cf26308-e979-456c-92a8-f2bdf0d6948f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.682121] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.685892] env[65107]: DEBUG nova.network.neutron [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [{"id": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "address": "fa:16:3e:c6:73:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd88e66c9-ba", "ovs_interfaceid": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.933533] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.916s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.934103] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1057.936946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.864s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.937088] env[65107]: DEBUG nova.objects.instance [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lazy-loading 'resources' on Instance uuid d67468c6-9aaf-49c4-afe4-5d9856c4af7a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.977749] env[65107]: DEBUG nova.compute.manager [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Received event network-changed-2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1057.978099] env[65107]: DEBUG nova.compute.manager [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Refreshing instance network info cache due to event network-changed-2cf26308-e979-456c-92a8-f2bdf0d6948f. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1057.978099] env[65107]: DEBUG oslo_concurrency.lockutils [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Acquiring lock "refresh_cache-a5ea4075-616a-49ce-aa93-f1db495a022b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.052999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-a5ea4075-616a-49ce-aa93-f1db495a022b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.053647] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Instance network_info: |[{"id": "2cf26308-e979-456c-92a8-f2bdf0d6948f", "address": "fa:16:3e:ae:d5:48", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf26308-e9", "ovs_interfaceid": "2cf26308-e979-456c-92a8-f2bdf0d6948f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1058.055302] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1058.058332] env[65107]: DEBUG oslo_concurrency.lockutils [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Acquired lock "refresh_cache-a5ea4075-616a-49ce-aa93-f1db495a022b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.058531] env[65107]: DEBUG nova.network.neutron [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Refreshing network info cache for port 2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1058.060273] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:d5:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cf26308-e979-456c-92a8-f2bdf0d6948f', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.068581] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1058.071964] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.072272] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b68bb9b-ef81-43d2-a9c7-2fa7adc143b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.096816] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1058.097121] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1058.097300] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1058.097482] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1058.097715] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1058.097882] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1058.098130] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.098315] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1058.098483] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1058.098678] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1058.098869] env[65107]: DEBUG nova.virt.hardware [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1058.100220] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca6e2f2-1d8f-4809-b48d-4c793ad442bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.104812] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.104812] env[65107]: value = "task-5103414" [ 1058.104812] env[65107]: _type = "Task" [ 1058.104812] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.112597] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8831d49e-5949-4643-a4f0-32a449dca133 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.122052] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103414, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.179358] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.189297] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.441044] env[65107]: DEBUG nova.compute.utils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1058.445506] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1058.445683] env[65107]: DEBUG nova.network.neutron [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1058.446222] env[65107]: WARNING neutronclient.v2_0.client [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.446311] env[65107]: WARNING neutronclient.v2_0.client [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.446962] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.447332] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.463158] env[65107]: DEBUG nova.network.neutron [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Successfully updated port: 9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1058.505234] env[65107]: DEBUG nova.policy [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d55cff40f12420b8320448083ca8f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7784762e36ae4c6fb30012ec801c3b88', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1058.570124] env[65107]: WARNING neutronclient.v2_0.client [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.570898] env[65107]: WARNING openstack [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.571256] env[65107]: WARNING openstack [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.620770] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103414, 'name': CreateVM_Task, 'duration_secs': 0.32781} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.625507] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.627837] env[65107]: WARNING neutronclient.v2_0.client [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.628416] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.628916] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.629365] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1058.629741] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e93b0e61-3b08-4ca3-9ce5-f9dbdfa50ba5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.637850] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1058.637850] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5236efdb-993b-5bf0-6ca8-8b9a77af0122" [ 1058.637850] env[65107]: _type = "Task" [ 1058.637850] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.655628] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5236efdb-993b-5bf0-6ca8-8b9a77af0122, 'name': SearchDatastore_Task, 'duration_secs': 0.011712} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.656091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.656443] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.656801] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.657183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.657326] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.657742] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2c9326d-0ad4-4924-acc6-22d28eb72b23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.680514] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.684355] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.684666] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.685510] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19eda4ca-1253-4bda-b189-f169335453ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.692605] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1058.692605] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525dba98-1222-41e2-0f41-89485b9398ac" [ 1058.692605] env[65107]: _type = "Task" [ 1058.692605] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.699270] env[65107]: WARNING openstack [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.699270] env[65107]: WARNING openstack [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.726197] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525dba98-1222-41e2-0f41-89485b9398ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.738093] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80439f9a-dc6b-493c-994f-fc29dc1f2f0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.767556] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb2bbda-a905-4c0b-9fba-44c3bdd43390 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.778909] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.786215] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371156aa-47c4-4a8a-922d-8eabdc63f5e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.795041] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eaca5f0-5fc0-48c9-bd70-ca57480d862f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.826852] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a83633-ac78-42fb-aa0c-bbe8874be0aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.830532] env[65107]: DEBUG nova.network.neutron [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Successfully created port: 6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1058.838952] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34ae663-31d2-48e8-9e10-68b12bddfd5f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.854468] env[65107]: DEBUG nova.compute.provider_tree [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.857855] env[65107]: WARNING neutronclient.v2_0.client [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.858539] env[65107]: WARNING openstack [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.858883] env[65107]: WARNING openstack [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.946515] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1058.966524] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.966817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.967093] env[65107]: DEBUG nova.network.neutron [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1058.994802] env[65107]: DEBUG nova.network.neutron [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Updated VIF entry in instance network info cache for port 2cf26308-e979-456c-92a8-f2bdf0d6948f. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1058.995258] env[65107]: DEBUG nova.network.neutron [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Updating instance_info_cache with network_info: [{"id": "2cf26308-e979-456c-92a8-f2bdf0d6948f", "address": "fa:16:3e:ae:d5:48", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf26308-e9", "ovs_interfaceid": "2cf26308-e979-456c-92a8-f2bdf0d6948f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.189891] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.206853] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525dba98-1222-41e2-0f41-89485b9398ac, 'name': SearchDatastore_Task, 'duration_secs': 0.031927} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.208189] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1012f15-c411-4017-906b-eaa0bda3453f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.217012] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1059.217012] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522deb01-578a-0e9e-371d-2dfaf44aa017" [ 1059.217012] env[65107]: _type = "Task" [ 1059.217012] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.227911] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522deb01-578a-0e9e-371d-2dfaf44aa017, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.286343] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1059.287071] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-423bdca2-0fa1-478a-8c9a-b296f3f33263 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.297025] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1059.297025] env[65107]: value = "task-5103415" [ 1059.297025] env[65107]: _type = "Task" [ 1059.297025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.310416] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103415, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.368766] env[65107]: DEBUG nova.scheduler.client.report [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1059.470841] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.471949] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.498836] env[65107]: DEBUG oslo_concurrency.lockutils [req-d28ff059-deeb-446e-9944-90238cba5e29 req-60d8440b-fced-4e74-b454-428daaa2dddd service nova] Releasing lock "refresh_cache-a5ea4075-616a-49ce-aa93-f1db495a022b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.588951] env[65107]: DEBUG nova.network.neutron [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1059.622215] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.622521] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.686835] env[65107]: DEBUG oslo_vmware.api [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103409, 'name': ReconfigVM_Task, 'duration_secs': 5.9287} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.687275] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.687524] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Reconfigured VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1059.688201] env[65107]: WARNING neutronclient.v2_0.client [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.688653] env[65107]: WARNING neutronclient.v2_0.client [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.689503] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.690011] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.704834] env[65107]: WARNING neutronclient.v2_0.client [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.705804] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.706293] env[65107]: WARNING openstack [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.729067] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]522deb01-578a-0e9e-371d-2dfaf44aa017, 'name': SearchDatastore_Task, 'duration_secs': 0.030907} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.729381] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.729660] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] a5ea4075-616a-49ce-aa93-f1db495a022b/a5ea4075-616a-49ce-aa93-f1db495a022b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.729950] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d9796b9-b25c-4988-ab91-1d6d0b92ca2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.737936] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1059.737936] env[65107]: value = "task-5103416" [ 1059.737936] env[65107]: _type = "Task" [ 1059.737936] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.748522] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103416, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.809427] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103415, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.917319] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.917319] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.925s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.917319] env[65107]: INFO nova.compute.claims [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.929532] env[65107]: INFO nova.scheduler.client.report [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Deleted allocations for instance d67468c6-9aaf-49c4-afe4-5d9856c4af7a [ 1059.963731] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1060.018693] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1060.018934] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1060.019845] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1060.020095] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1060.020676] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1060.020676] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1060.021489] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.021489] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1060.021489] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1060.021752] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1060.021974] env[65107]: DEBUG nova.virt.hardware [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1060.023542] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc7f6ce-745a-4476-9406-bb144287061c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.041563] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d8a622-4eb5-41ae-a806-08246adc8767 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.239044] env[65107]: DEBUG nova.network.neutron [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [{"id": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "address": "fa:16:3e:44:a8:f3", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e9a520d-89", "ovs_interfaceid": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1060.254887] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103416, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.314592] env[65107]: DEBUG oslo_vmware.api [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103415, 'name': PowerOnVM_Task, 'duration_secs': 0.661873} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.316348] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.316348] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d2ffd8-fb8e-4d2b-b403-288109b97a61 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance '00574b77-dad6-4f0a-bbcc-20a2b4a5df65' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1060.383114] env[65107]: DEBUG nova.compute.manager [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Received event network-vif-plugged-9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1060.383364] env[65107]: DEBUG oslo_concurrency.lockutils [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Acquiring lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.383570] env[65107]: DEBUG oslo_concurrency.lockutils [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.383800] env[65107]: DEBUG oslo_concurrency.lockutils [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.383909] env[65107]: DEBUG nova.compute.manager [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] No waiting events found dispatching network-vif-plugged-9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1060.385374] env[65107]: WARNING nova.compute.manager [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Received unexpected event network-vif-plugged-9e9a520d-8984-4d4d-934b-f46dbdac5f4d for instance with vm_state building and task_state spawning. [ 1060.385744] env[65107]: DEBUG nova.compute.manager [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Received event network-changed-9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1060.386037] env[65107]: DEBUG nova.compute.manager [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Refreshing instance network info cache due to event network-changed-9e9a520d-8984-4d4d-934b-f46dbdac5f4d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1060.386383] env[65107]: DEBUG oslo_concurrency.lockutils [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Acquiring lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.444023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-26adac00-26f6-4697-ab16-23f8461dd1d0 tempest-ImagesTestJSON-2026451637 tempest-ImagesTestJSON-2026451637-project-member] Lock "d67468c6-9aaf-49c4-afe4-5d9856c4af7a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.678s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.517920] env[65107]: DEBUG nova.network.neutron [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Successfully updated port: 6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1060.693117] env[65107]: DEBUG nova.compute.manager [req-03f45e2f-8f89-4fc2-9519-24bc3da26ec3 req-b316aaa2-4ced-45f0-adaa-8a17eb2fd6d9 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Received event network-vif-plugged-6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1060.693117] env[65107]: DEBUG oslo_concurrency.lockutils [req-03f45e2f-8f89-4fc2-9519-24bc3da26ec3 req-b316aaa2-4ced-45f0-adaa-8a17eb2fd6d9 service nova] Acquiring lock "8dbed811-3d4e-4cae-9981-0334801013aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.693117] env[65107]: DEBUG oslo_concurrency.lockutils [req-03f45e2f-8f89-4fc2-9519-24bc3da26ec3 req-b316aaa2-4ced-45f0-adaa-8a17eb2fd6d9 service nova] Lock "8dbed811-3d4e-4cae-9981-0334801013aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.693117] env[65107]: DEBUG oslo_concurrency.lockutils [req-03f45e2f-8f89-4fc2-9519-24bc3da26ec3 req-b316aaa2-4ced-45f0-adaa-8a17eb2fd6d9 service nova] Lock "8dbed811-3d4e-4cae-9981-0334801013aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.693117] env[65107]: DEBUG nova.compute.manager [req-03f45e2f-8f89-4fc2-9519-24bc3da26ec3 req-b316aaa2-4ced-45f0-adaa-8a17eb2fd6d9 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] No waiting events found dispatching network-vif-plugged-6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1060.693117] env[65107]: WARNING nova.compute.manager [req-03f45e2f-8f89-4fc2-9519-24bc3da26ec3 req-b316aaa2-4ced-45f0-adaa-8a17eb2fd6d9 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Received unexpected event network-vif-plugged-6091d1cb-29b9-4a22-8827-990e8ea7d318 for instance with vm_state building and task_state spawning. [ 1060.747448] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.747448] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Instance network_info: |[{"id": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "address": "fa:16:3e:44:a8:f3", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e9a520d-89", "ovs_interfaceid": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1060.747448] env[65107]: DEBUG oslo_concurrency.lockutils [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Acquired lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.747878] env[65107]: DEBUG nova.network.neutron [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Refreshing network info cache for port 9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1060.749443] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:a8:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e9a520d-8984-4d4d-934b-f46dbdac5f4d', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.762295] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1060.764525] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1060.764940] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c1798f5-ff68-4610-8149-2fe9d66be9b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.786353] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103416, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.794267] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.794267] env[65107]: value = "task-5103417" [ 1060.794267] env[65107]: _type = "Task" [ 1060.794267] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.808030] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103417, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.020734] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-8dbed811-3d4e-4cae-9981-0334801013aa" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.020980] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-8dbed811-3d4e-4cae-9981-0334801013aa" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.021179] env[65107]: DEBUG nova.network.neutron [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1061.174935] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.174935] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.174935] env[65107]: DEBUG nova.network.neutron [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1061.187788] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a0e09f-519f-4445-b0aa-da928892fc27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.197979] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5660be-2034-4972-b1b5-514840391bdc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.241724] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3107a7d5-2215-47d9-8e4c-40eabbe3f933 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.256499] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620d0fbc-eb68-40f7-b93e-2fca32bd6a15 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.260582] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103416, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.264212] env[65107]: WARNING neutronclient.v2_0.client [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.264906] env[65107]: WARNING openstack [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.265280] env[65107]: WARNING openstack [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.281356] env[65107]: DEBUG nova.compute.provider_tree [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.305487] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103417, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.401564] env[65107]: WARNING openstack [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.401725] env[65107]: WARNING openstack [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.522134] env[65107]: WARNING neutronclient.v2_0.client [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.523259] env[65107]: WARNING openstack [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.523862] env[65107]: WARNING openstack [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.533662] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.534228] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.623281] env[65107]: DEBUG nova.network.neutron [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1061.660999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.661491] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.661761] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.662112] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.662424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.665818] env[65107]: INFO nova.compute.manager [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Terminating instance [ 1061.678273] env[65107]: WARNING neutronclient.v2_0.client [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.679317] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.679853] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.724416] env[65107]: DEBUG nova.network.neutron [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updated VIF entry in instance network info cache for port 9e9a520d-8984-4d4d-934b-f46dbdac5f4d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1061.724801] env[65107]: DEBUG nova.network.neutron [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [{"id": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "address": "fa:16:3e:44:a8:f3", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e9a520d-89", "ovs_interfaceid": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1061.738644] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.739080] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.756681] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103416, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.587615} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.756975] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] a5ea4075-616a-49ce-aa93-f1db495a022b/a5ea4075-616a-49ce-aa93-f1db495a022b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.757206] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.757460] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49618c2c-8574-4608-9beb-9117edda72a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.764569] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1061.764569] env[65107]: value = "task-5103418" [ 1061.764569] env[65107]: _type = "Task" [ 1061.764569] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.773591] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103418, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.784768] env[65107]: DEBUG nova.scheduler.client.report [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1061.809591] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103417, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.961501] env[65107]: WARNING neutronclient.v2_0.client [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.962586] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.963152] env[65107]: WARNING openstack [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.987151] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.989117] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.139964] env[65107]: DEBUG nova.network.neutron [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Updating instance_info_cache with network_info: [{"id": "6091d1cb-29b9-4a22-8827-990e8ea7d318", "address": "fa:16:3e:64:c8:04", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6091d1cb-29", "ovs_interfaceid": "6091d1cb-29b9-4a22-8827-990e8ea7d318", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.172937] env[65107]: DEBUG nova.compute.manager [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1062.173051] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1062.183181] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c50de2-0725-4575-ae40-4e2b63bf8f2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.190653] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.190992] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5db351a-162c-4265-a3e4-73e5300df390 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.198650] env[65107]: DEBUG oslo_vmware.api [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1062.198650] env[65107]: value = "task-5103419" [ 1062.198650] env[65107]: _type = "Task" [ 1062.198650] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.204776] env[65107]: WARNING neutronclient.v2_0.client [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.205637] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.206090] env[65107]: WARNING openstack [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.218317] env[65107]: DEBUG oslo_vmware.api [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103419, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.228022] env[65107]: DEBUG oslo_concurrency.lockutils [req-bff08d77-db06-4f7e-9dcd-b72c40fbcd6b req-38867b46-8daf-4b47-b99e-36c6ea98165e service nova] Releasing lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.279326] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103418, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.230505} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.279654] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.280562] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ccb99a9-af6d-44f2-8ad6-c8b1d5733b4f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.297821] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.298382] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1062.311231] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] a5ea4075-616a-49ce-aa93-f1db495a022b/a5ea4075-616a-49ce-aa93-f1db495a022b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.311991] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.844s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.313349] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f890b59-2ded-4f1b-9480-534fd9268faa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.337967] env[65107]: INFO nova.network.neutron [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Port fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1062.338136] env[65107]: INFO nova.network.neutron [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Port 57c8daf0-4e71-4f3e-8e92-cde14428fbe3 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1062.338562] env[65107]: DEBUG nova.network.neutron [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [{"id": "e5280424-9c40-4f31-a39e-260899363c1d", "address": "fa:16:3e:dc:63:0d", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5280424-9c", "ovs_interfaceid": "e5280424-9c40-4f31-a39e-260899363c1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.342334] env[65107]: INFO nova.compute.claims [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.359625] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103417, 'name': CreateVM_Task, 'duration_secs': 1.399766} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.361520] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1062.362352] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1062.362352] env[65107]: value = "task-5103420" [ 1062.362352] env[65107]: _type = "Task" [ 1062.362352] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.363069] env[65107]: WARNING neutronclient.v2_0.client [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.363592] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.364234] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1062.364773] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1062.365422] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-743447b2-7c23-4f60-856b-f3bcfcc57907 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.376375] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1062.376375] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1b8e7-2fac-1b05-d9e3-8eb801b9be5e" [ 1062.376375] env[65107]: _type = "Task" [ 1062.376375] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.380482] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.391503] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1b8e7-2fac-1b05-d9e3-8eb801b9be5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.428651] env[65107]: DEBUG nova.compute.manager [req-b36fc488-b48b-4593-b975-3cd505bd11b9 req-18c5dc39-3c3d-462a-86e5-a68590ccfb5e service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-vif-deleted-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1062.643266] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-8dbed811-3d4e-4cae-9981-0334801013aa" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.643654] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance network_info: |[{"id": "6091d1cb-29b9-4a22-8827-990e8ea7d318", "address": "fa:16:3e:64:c8:04", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6091d1cb-29", "ovs_interfaceid": "6091d1cb-29b9-4a22-8827-990e8ea7d318", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1062.644152] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:c8:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6091d1cb-29b9-4a22-8827-990e8ea7d318', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.652022] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1062.652267] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1062.652512] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0320238d-77c5-418a-b859-623bd1a40d65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.677547] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.677547] env[65107]: value = "task-5103421" [ 1062.677547] env[65107]: _type = "Task" [ 1062.677547] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.687665] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103421, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.709787] env[65107]: DEBUG oslo_vmware.api [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103419, 'name': PowerOffVM_Task, 'duration_secs': 0.224682} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.709787] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1062.709999] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1062.710108] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8e36146-4b4a-4b38-9233-b3564dd23eec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.744531] env[65107]: DEBUG nova.compute.manager [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Received event network-changed-6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1062.745098] env[65107]: DEBUG nova.compute.manager [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Refreshing instance network info cache due to event network-changed-6091d1cb-29b9-4a22-8827-990e8ea7d318. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1062.745098] env[65107]: DEBUG oslo_concurrency.lockutils [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Acquiring lock "refresh_cache-8dbed811-3d4e-4cae-9981-0334801013aa" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.745428] env[65107]: DEBUG oslo_concurrency.lockutils [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Acquired lock "refresh_cache-8dbed811-3d4e-4cae-9981-0334801013aa" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1062.745639] env[65107]: DEBUG nova.network.neutron [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Refreshing network info cache for port 6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1062.814364] env[65107]: DEBUG nova.compute.utils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1062.820026] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1062.820026] env[65107]: DEBUG nova.network.neutron [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1062.820026] env[65107]: WARNING neutronclient.v2_0.client [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.820026] env[65107]: WARNING neutronclient.v2_0.client [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.820026] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.820026] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.838586] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1062.838861] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1062.839074] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleting the datastore file [datastore1] afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1062.839812] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0614d261-4248-46d7-a18c-f9f741f02950 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.847524] env[65107]: DEBUG oslo_vmware.api [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1062.847524] env[65107]: value = "task-5103423" [ 1062.847524] env[65107]: _type = "Task" [ 1062.847524] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.853498] env[65107]: INFO nova.compute.resource_tracker [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating resource usage from migration f9d5050a-6b81-4c45-ae1d-c9b5fa915cc2 [ 1062.856280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-afa1a4cc-aa43-4e10-982f-edaf5cdee584" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.865737] env[65107]: DEBUG oslo_vmware.api [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.876990] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.890332] env[65107]: DEBUG nova.policy [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a2861bd2b8a40b4acdb9c79142542db', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '741a5435884941de99a9983f2c8cc560', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1062.900741] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a1b8e7-2fac-1b05-d9e3-8eb801b9be5e, 'name': SearchDatastore_Task, 'duration_secs': 0.016315} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.901496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.901782] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1062.902107] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.902289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1062.902515] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1062.902832] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39b60c76-1357-4487-b7d4-185a322acde4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.922999] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.923232] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1062.924094] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ceaf29db-e858-42a6-bb2f-90bcda80fa14 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.931155] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1062.931155] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523ce1ae-aa6f-ce85-9d0c-6140cb8c1221" [ 1062.931155] env[65107]: _type = "Task" [ 1062.931155] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.943431] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523ce1ae-aa6f-ce85-9d0c-6140cb8c1221, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.165267] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79e59ac-2f6e-44b0-816a-b4a7d7d9dbff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.186055] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017e25cc-8410-4fb8-9d66-d18e08387993 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.195813] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103421, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.224313] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a036d2-83df-4e89-bc92-deb04450dadb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.233945] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be64a5ab-a980-4cc5-b4ad-4a8da6032d34 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.252267] env[65107]: WARNING neutronclient.v2_0.client [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.253659] env[65107]: WARNING openstack [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1063.253659] env[65107]: WARNING openstack [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1063.261150] env[65107]: DEBUG nova.compute.provider_tree [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.311886] env[65107]: DEBUG nova.network.neutron [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Successfully created port: 32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1063.314436] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.314774] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.314967] env[65107]: DEBUG nova.compute.manager [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Going to confirm migration 4 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1063.318321] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1063.361629] env[65107]: DEBUG oslo_concurrency.lockutils [None req-039ff3e5-b311-445b-85a1-560b9ea3c0c6 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-afa1a4cc-aa43-4e10-982f-edaf5cdee584-fbe601d0-6bb2-4d9d-b98a-8d174aa4a6f4" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.301s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.363031] env[65107]: DEBUG oslo_vmware.api [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281907} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.366507] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1063.366507] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1063.366618] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1063.367193] env[65107]: INFO nova.compute.manager [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1063.367193] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1063.368474] env[65107]: DEBUG nova.compute.manager [-] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1063.369028] env[65107]: DEBUG nova.network.neutron [-] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1063.369028] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.369411] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1063.369670] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1063.388625] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103420, 'name': ReconfigVM_Task, 'duration_secs': 1.001852} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.388936] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Reconfigured VM instance instance-00000060 to attach disk [datastore1] a5ea4075-616a-49ce-aa93-f1db495a022b/a5ea4075-616a-49ce-aa93-f1db495a022b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.389633] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37711ca9-0de7-4ee8-b88b-eeddf1595e19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.397490] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1063.397490] env[65107]: value = "task-5103424" [ 1063.397490] env[65107]: _type = "Task" [ 1063.397490] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.407502] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103424, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.443345] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523ce1ae-aa6f-ce85-9d0c-6140cb8c1221, 'name': SearchDatastore_Task, 'duration_secs': 0.025269} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.444208] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-396c8a13-81cb-4371-aad1-96a1d4241474 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.456305] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1063.456305] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b6aae-2dc0-1d0e-3762-c4ba0418213e" [ 1063.456305] env[65107]: _type = "Task" [ 1063.456305] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.466119] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525b6aae-2dc0-1d0e-3762-c4ba0418213e, 'name': SearchDatastore_Task, 'duration_secs': 0.012301} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.466390] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.466677] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b/7373fbbe-52c1-4955-8da9-a5563d71391b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1063.466952] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21e332c1-b587-4a1e-9ff4-b7e29f43f5b4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.475137] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1063.475137] env[65107]: value = "task-5103425" [ 1063.475137] env[65107]: _type = "Task" [ 1063.475137] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.483729] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.509812] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.525033] env[65107]: WARNING openstack [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1063.525480] env[65107]: WARNING openstack [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1063.568678] env[65107]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 57c8daf0-4e71-4f3e-8e92-cde14428fbe3 could not be found.", "detail": ""}} {{(pid=65107) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:265}} [ 1063.569159] env[65107]: DEBUG nova.network.neutron [-] Unable to show port 57c8daf0-4e71-4f3e-8e92-cde14428fbe3 as it no longer exists. {{(pid=65107) _unbind_ports /opt/stack/nova/nova/network/neutron.py:700}} [ 1063.691642] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103421, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.763917] env[65107]: DEBUG nova.scheduler.client.report [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1063.820667] env[65107]: WARNING neutronclient.v2_0.client [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.908423] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103424, 'name': Rename_Task, 'duration_secs': 0.154732} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.908723] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.909070] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32eba2ac-9150-4a51-8995-e102a9a9b342 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.916286] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1063.916286] env[65107]: value = "task-5103426" [ 1063.916286] env[65107]: _type = "Task" [ 1063.916286] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.925079] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103426, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.984982] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.052124] env[65107]: WARNING neutronclient.v2_0.client [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.052850] env[65107]: WARNING openstack [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.053260] env[65107]: WARNING openstack [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.191015] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103421, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.269592] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 1.957s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1064.269592] env[65107]: INFO nova.compute.manager [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Migrating [ 1064.329035] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1064.362568] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1064.362733] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1064.362840] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1064.363156] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1064.363579] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1064.363579] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1064.363969] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.364509] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1064.364613] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1064.364851] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1064.365108] env[65107]: DEBUG nova.virt.hardware [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1064.366924] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbe2918-5743-48af-803e-77c4c790e0be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.378514] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906e6534-cc2f-45c2-9680-75bcac5645a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.430039] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103426, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.485319] env[65107]: WARNING neutronclient.v2_0.client [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.485733] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.485890] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.486061] env[65107]: DEBUG nova.network.neutron [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1064.486260] env[65107]: DEBUG nova.objects.instance [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'info_cache' on Instance uuid 00574b77-dad6-4f0a-bbcc-20a2b4a5df65 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.491750] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103425, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.525363] env[65107]: DEBUG nova.network.neutron [-] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1064.694088] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103421, 'name': CreateVM_Task, 'duration_secs': 1.561266} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.694462] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.694961] env[65107]: WARNING neutronclient.v2_0.client [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.695768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.695768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.695768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1064.696226] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeaa9074-72bc-460a-ba30-e9cf4f9befd3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.702028] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1064.702028] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5237229c-6ecc-3db5-8c33-6f986be2b78c" [ 1064.702028] env[65107]: _type = "Task" [ 1064.702028] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.712650] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5237229c-6ecc-3db5-8c33-6f986be2b78c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.787033] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.787299] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.787427] env[65107]: DEBUG nova.network.neutron [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1064.821675] env[65107]: DEBUG nova.network.neutron [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Updated VIF entry in instance network info cache for port 6091d1cb-29b9-4a22-8827-990e8ea7d318. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1064.822062] env[65107]: DEBUG nova.network.neutron [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Updating instance_info_cache with network_info: [{"id": "6091d1cb-29b9-4a22-8827-990e8ea7d318", "address": "fa:16:3e:64:c8:04", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6091d1cb-29", "ovs_interfaceid": "6091d1cb-29b9-4a22-8827-990e8ea7d318", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1064.911777] env[65107]: DEBUG nova.network.neutron [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Successfully updated port: 32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1064.928801] env[65107]: DEBUG oslo_vmware.api [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103426, 'name': PowerOnVM_Task, 'duration_secs': 0.621799} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.930254] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.930518] env[65107]: INFO nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Took 9.31 seconds to spawn the instance on the hypervisor. [ 1064.930680] env[65107]: DEBUG nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1064.931767] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d575d8c-f173-4b9f-a853-48d9de6eabc8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.987608] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103425, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.052687} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.987879] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b/7373fbbe-52c1-4955-8da9-a5563d71391b.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1064.988249] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.988361] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fbfd835-31b3-4e52-85b2-bee1a2b97ec7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.996066] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1064.996066] env[65107]: value = "task-5103427" [ 1064.996066] env[65107]: _type = "Task" [ 1064.996066] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.005561] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.029461] env[65107]: INFO nova.compute.manager [-] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Took 1.66 seconds to deallocate network for instance. [ 1065.214688] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5237229c-6ecc-3db5-8c33-6f986be2b78c, 'name': SearchDatastore_Task, 'duration_secs': 0.013873} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.214964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.215223] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1065.215467] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.215670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1065.215759] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1065.216034] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9444937-5095-4e3d-83cb-ed2aa035dd36 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.226445] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1065.226618] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1065.227424] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3abd7947-bb7b-4bae-bfb1-108061676b1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.233338] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1065.233338] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f07c46-5b31-d7bb-2944-f2e21440a0a7" [ 1065.233338] env[65107]: _type = "Task" [ 1065.233338] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.243789] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f07c46-5b31-d7bb-2944-f2e21440a0a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.266122] env[65107]: DEBUG nova.compute.manager [req-8b64d613-0cf5-4827-93c9-4e783fb9dab6 req-7c93b9d6-3427-4380-990c-255f5f9c7d3e service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-vif-deleted-e5280424-9c40-4f31-a39e-260899363c1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1065.291104] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.291872] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.292249] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.301429] env[65107]: DEBUG nova.compute.manager [req-eed534ef-7542-4734-8f3e-2d1832ab7939 req-5466bd3f-3b92-4537-b56d-713bebdef604 service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Received event network-vif-plugged-32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1065.301971] env[65107]: DEBUG oslo_concurrency.lockutils [req-eed534ef-7542-4734-8f3e-2d1832ab7939 req-5466bd3f-3b92-4537-b56d-713bebdef604 service nova] Acquiring lock "f6536493-fb80-4fb1-9ee1-e530734e9adf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.302204] env[65107]: DEBUG oslo_concurrency.lockutils [req-eed534ef-7542-4734-8f3e-2d1832ab7939 req-5466bd3f-3b92-4537-b56d-713bebdef604 service nova] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.302648] env[65107]: DEBUG oslo_concurrency.lockutils [req-eed534ef-7542-4734-8f3e-2d1832ab7939 req-5466bd3f-3b92-4537-b56d-713bebdef604 service nova] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.302648] env[65107]: DEBUG nova.compute.manager [req-eed534ef-7542-4734-8f3e-2d1832ab7939 req-5466bd3f-3b92-4537-b56d-713bebdef604 service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] No waiting events found dispatching network-vif-plugged-32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1065.303040] env[65107]: WARNING nova.compute.manager [req-eed534ef-7542-4734-8f3e-2d1832ab7939 req-5466bd3f-3b92-4537-b56d-713bebdef604 service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Received unexpected event network-vif-plugged-32156605-37d7-4888-9a96-3ac243961051 for instance with vm_state building and task_state spawning. [ 1065.324699] env[65107]: DEBUG oslo_concurrency.lockutils [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] Releasing lock "refresh_cache-8dbed811-3d4e-4cae-9981-0334801013aa" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.324995] env[65107]: DEBUG nova.compute.manager [req-694d58c8-5443-471e-b74a-6149dec762a7 req-85bb69da-f238-41bf-b148-f49d5a133e44 service nova] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Received event network-vif-deleted-57c8daf0-4e71-4f3e-8e92-cde14428fbe3 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1065.418103] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.418535] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.425529] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "refresh_cache-f6536493-fb80-4fb1-9ee1-e530734e9adf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.425711] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquired lock "refresh_cache-f6536493-fb80-4fb1-9ee1-e530734e9adf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1065.425868] env[65107]: DEBUG nova.network.neutron [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1065.456229] env[65107]: INFO nova.compute.manager [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Took 22.19 seconds to build instance. [ 1065.485223] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.485825] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.486193] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.497644] env[65107]: WARNING neutronclient.v2_0.client [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.498414] env[65107]: WARNING openstack [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.498757] env[65107]: WARNING openstack [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.517833] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073949} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.518101] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.518921] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4b30ef-2b1a-4a1d-a57e-e4aaa07b806b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.537092] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.537797] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.538099] env[65107]: DEBUG nova.objects.instance [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'resources' on Instance uuid afa1a4cc-aa43-4e10-982f-edaf5cdee584 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.548311] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b/7373fbbe-52c1-4955-8da9-a5563d71391b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.558389] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6675cd87-8f9f-485e-9c09-cb96d5e349e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.582027] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1065.582027] env[65107]: value = "task-5103428" [ 1065.582027] env[65107]: _type = "Task" [ 1065.582027] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.592742] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.623512] env[65107]: DEBUG nova.network.neutron [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1065.683236] env[65107]: WARNING openstack [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.683921] env[65107]: WARNING openstack [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.744442] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f07c46-5b31-d7bb-2944-f2e21440a0a7, 'name': SearchDatastore_Task, 'duration_secs': 0.01026} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.749030] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14126e9-259a-44c4-8f6e-cb05ba21e950 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.749603] env[65107]: WARNING neutronclient.v2_0.client [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.750126] env[65107]: WARNING openstack [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.750468] env[65107]: WARNING openstack [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.759377] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1065.759377] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52637ee5-9449-8642-b679-f79233869d16" [ 1065.759377] env[65107]: _type = "Task" [ 1065.759377] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.773391] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52637ee5-9449-8642-b679-f79233869d16, 'name': SearchDatastore_Task, 'duration_secs': 0.011554} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.776290] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.776696] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.778680] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b70aea6-f3ba-4b29-b260-cb7c88e59512 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.785400] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1065.785400] env[65107]: value = "task-5103429" [ 1065.785400] env[65107]: _type = "Task" [ 1065.785400] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.794649] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103429, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.873734] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7119feed-d411-4f97-b459-506ab6dd11e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.882449] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fc908a-37e3-4b2b-9377-76567c981bdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.921746] env[65107]: DEBUG nova.network.neutron [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [{"id": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "address": "fa:16:3e:c6:73:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd88e66c9-ba", "ovs_interfaceid": "d88e66c9-ba7c-452d-8e44-8c9603f2b258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1065.926021] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3635e640-5af8-4cc6-95cd-29832548f917 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.929690] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.930104] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.946493] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568d4e06-de32-4303-a4c4-a264af6dc17d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.968538] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9275bcfe-9d18-410b-ba64-5d4843b1c585 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.713s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.969555] env[65107]: DEBUG nova.compute.provider_tree [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.004407] env[65107]: DEBUG nova.network.neutron [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1066.030122] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1066.030543] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1066.095925] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103428, 'name': ReconfigVM_Task, 'duration_secs': 0.421388} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.096253] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b/7373fbbe-52c1-4955-8da9-a5563d71391b.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.097028] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44954bc3-13dd-4be3-9730-7735c6086d05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.102793] env[65107]: WARNING neutronclient.v2_0.client [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1066.103282] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1066.103834] env[65107]: WARNING openstack [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1066.117513] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1066.117513] env[65107]: value = "task-5103430" [ 1066.117513] env[65107]: _type = "Task" [ 1066.117513] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.126331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.134728] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103430, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.297651] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103429, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49952} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.297939] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1066.298168] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.299210] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30b1c5c7-734a-48a7-b064-32ce697f9911 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.304583] env[65107]: DEBUG nova.network.neutron [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Updating instance_info_cache with network_info: [{"id": "32156605-37d7-4888-9a96-3ac243961051", "address": "fa:16:3e:b8:39:65", "network": {"id": "05cb028c-3182-4757-9ff8-0e7e3fb803e8", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2014677764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741a5435884941de99a9983f2c8cc560", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32156605-37", "ovs_interfaceid": "32156605-37d7-4888-9a96-3ac243961051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1066.310129] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1066.310129] env[65107]: value = "task-5103431" [ 1066.310129] env[65107]: _type = "Task" [ 1066.310129] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.320974] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.430286] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-00574b77-dad6-4f0a-bbcc-20a2b4a5df65" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.430608] env[65107]: DEBUG nova.objects.instance [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'migration_context' on Instance uuid 00574b77-dad6-4f0a-bbcc-20a2b4a5df65 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.473847] env[65107]: DEBUG nova.scheduler.client.report [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1066.629840] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103430, 'name': Rename_Task, 'duration_secs': 0.294043} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.630139] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1066.630411] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-725888e1-4c0e-4944-9f0e-e917dd6633d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.643310] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1066.643310] env[65107]: value = "task-5103432" [ 1066.643310] env[65107]: _type = "Task" [ 1066.643310] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.652698] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.805475] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Releasing lock "refresh_cache-f6536493-fb80-4fb1-9ee1-e530734e9adf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.805882] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Instance network_info: |[{"id": "32156605-37d7-4888-9a96-3ac243961051", "address": "fa:16:3e:b8:39:65", "network": {"id": "05cb028c-3182-4757-9ff8-0e7e3fb803e8", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2014677764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741a5435884941de99a9983f2c8cc560", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32156605-37", "ovs_interfaceid": "32156605-37d7-4888-9a96-3ac243961051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1066.806523] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:39:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00a15667-7ca5-4dc9-be92-164750d87988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32156605-37d7-4888-9a96-3ac243961051', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.814962] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Creating folder: Project (741a5435884941de99a9983f2c8cc560). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1066.815875] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8199980-35fa-400a-8e67-666bf3ffba0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.828957] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09328} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.828957] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.829432] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ab6055-37d4-48e2-8391-3772760b1dce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.833912] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Created folder: Project (741a5435884941de99a9983f2c8cc560) in parent group-v992574. [ 1066.834043] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Creating folder: Instances. Parent ref: group-v992852. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1066.835049] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-864a50f5-547d-49ec-a43d-57ec9e7a178c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.856884] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.858464] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8d66db1-3afe-45b8-86fa-d4aade8f967e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.873558] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Created folder: Instances in parent group-v992852. [ 1066.873823] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1066.874048] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.874764] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12b86a2b-274c-43cc-b1f3-e04e39ea138d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.894306] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1066.894306] env[65107]: value = "task-5103435" [ 1066.894306] env[65107]: _type = "Task" [ 1066.894306] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.899951] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.899951] env[65107]: value = "task-5103436" [ 1066.899951] env[65107]: _type = "Task" [ 1066.899951] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.909140] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.915665] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103436, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.933894] env[65107]: DEBUG nova.objects.base [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Object Instance<00574b77-dad6-4f0a-bbcc-20a2b4a5df65> lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1066.935327] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1246055-ed61-4f4d-8b04-1378cbd6c2ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.958699] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66b4def3-2f19-42bb-95eb-b84318f27301 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.966526] env[65107]: DEBUG oslo_vmware.api [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1066.966526] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207c4d5-ea21-9dcd-df9c-02d7ae855263" [ 1066.966526] env[65107]: _type = "Task" [ 1066.966526] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.976999] env[65107]: DEBUG oslo_vmware.api [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207c4d5-ea21-9dcd-df9c-02d7ae855263, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.980113] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.443s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.019766] env[65107]: INFO nova.scheduler.client.report [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted allocations for instance afa1a4cc-aa43-4e10-982f-edaf5cdee584 [ 1067.116033] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "a5ea4075-616a-49ce-aa93-f1db495a022b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.116438] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.116729] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "a5ea4075-616a-49ce-aa93-f1db495a022b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.117043] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.117294] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.120934] env[65107]: INFO nova.compute.manager [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Terminating instance [ 1067.157421] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103432, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.340706] env[65107]: DEBUG nova.compute.manager [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Received event network-changed-32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1067.341036] env[65107]: DEBUG nova.compute.manager [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Refreshing instance network info cache due to event network-changed-32156605-37d7-4888-9a96-3ac243961051. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1067.341362] env[65107]: DEBUG oslo_concurrency.lockutils [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Acquiring lock "refresh_cache-f6536493-fb80-4fb1-9ee1-e530734e9adf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.341478] env[65107]: DEBUG oslo_concurrency.lockutils [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Acquired lock "refresh_cache-f6536493-fb80-4fb1-9ee1-e530734e9adf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.341645] env[65107]: DEBUG nova.network.neutron [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Refreshing network info cache for port 32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1067.407101] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103435, 'name': ReconfigVM_Task, 'duration_secs': 0.412469} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.407833] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.408517] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd6ea843-f816-4393-bb82-3a63cc215276 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.413545] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103436, 'name': CreateVM_Task, 'duration_secs': 0.466508} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.414061] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.414589] env[65107]: WARNING neutronclient.v2_0.client [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.414945] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.415128] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.415442] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1067.415689] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec2d0ed3-f7bd-4bf4-9b0b-3f2cbcc78610 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.419220] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1067.419220] env[65107]: value = "task-5103437" [ 1067.419220] env[65107]: _type = "Task" [ 1067.419220] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.425979] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1067.425979] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5275cf14-1f6c-dc91-8248-58298cb64cb0" [ 1067.425979] env[65107]: _type = "Task" [ 1067.425979] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.432749] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103437, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.438974] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5275cf14-1f6c-dc91-8248-58298cb64cb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.479095] env[65107]: DEBUG oslo_vmware.api [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207c4d5-ea21-9dcd-df9c-02d7ae855263, 'name': SearchDatastore_Task, 'duration_secs': 0.008924} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.479479] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.479679] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.536049] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7311433b-1af7-4726-a807-26ecda311f12 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "afa1a4cc-aa43-4e10-982f-edaf5cdee584" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.874s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.626651] env[65107]: DEBUG nova.compute.manager [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1067.627021] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.628204] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0316c3a9-94ec-4bf6-bb47-aeb16516628a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.637349] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.637650] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-715b852e-93f4-4e0b-87d3-4c445df03429 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.647530] env[65107]: DEBUG oslo_vmware.api [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1067.647530] env[65107]: value = "task-5103438" [ 1067.647530] env[65107]: _type = "Task" [ 1067.647530] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.653035] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3933b119-3d18-4a83-84cb-e5cd8ae80f77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.688553] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.688806] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.690066] env[65107]: DEBUG oslo_vmware.api [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103438, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.690420] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.695076] env[65107]: DEBUG oslo_vmware.api [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103432, 'name': PowerOnVM_Task, 'duration_secs': 0.546123} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.695445] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.695646] env[65107]: INFO nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Took 9.64 seconds to spawn the instance on the hypervisor. [ 1067.695825] env[65107]: DEBUG nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1067.696811] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6aa8bc-2346-4ad9-b601-93c34a0b22be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.845330] env[65107]: WARNING neutronclient.v2_0.client [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.846726] env[65107]: WARNING openstack [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.847747] env[65107]: WARNING openstack [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.933952] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103437, 'name': Rename_Task, 'duration_secs': 0.345632} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.934841] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.935123] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2ded4ff-4303-404a-858a-908fd2f8abb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.940128] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5275cf14-1f6c-dc91-8248-58298cb64cb0, 'name': SearchDatastore_Task, 'duration_secs': 0.017862} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.940756] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1067.940980] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.941222] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.941368] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.941544] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.941851] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce6b7e11-e7b5-4aa4-8537-79b5af51f6fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.946111] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1067.946111] env[65107]: value = "task-5103439" [ 1067.946111] env[65107]: _type = "Task" [ 1067.946111] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.952487] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.952678] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.956582] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77c3752e-5ac1-4916-87cf-56bd492fb04d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.959418] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.963309] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1067.963309] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52088bc8-8b78-70e6-1064-5eaad3052e0a" [ 1067.963309] env[65107]: _type = "Task" [ 1067.963309] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.973168] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52088bc8-8b78-70e6-1064-5eaad3052e0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.024300] env[65107]: WARNING openstack [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.024854] env[65107]: WARNING openstack [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.100309] env[65107]: WARNING neutronclient.v2_0.client [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.100968] env[65107]: WARNING openstack [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.101476] env[65107]: WARNING openstack [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.165044] env[65107]: DEBUG oslo_vmware.api [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103438, 'name': PowerOffVM_Task, 'duration_secs': 0.221944} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.168298] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.168588] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.169264] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1c535ad-1fb3-49ae-ad08-92a210f97a61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.197627] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.198135] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1068.201820] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e25bec32-902b-476b-8e17-0ddfaa4f31ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.209697] env[65107]: DEBUG nova.network.neutron [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Updated VIF entry in instance network info cache for port 32156605-37d7-4888-9a96-3ac243961051. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1068.210131] env[65107]: DEBUG nova.network.neutron [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Updating instance_info_cache with network_info: [{"id": "32156605-37d7-4888-9a96-3ac243961051", "address": "fa:16:3e:b8:39:65", "network": {"id": "05cb028c-3182-4757-9ff8-0e7e3fb803e8", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2014677764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741a5435884941de99a9983f2c8cc560", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32156605-37", "ovs_interfaceid": "32156605-37d7-4888-9a96-3ac243961051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1068.220460] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1068.220460] env[65107]: value = "task-5103441" [ 1068.220460] env[65107]: _type = "Task" [ 1068.220460] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.228343] env[65107]: INFO nova.compute.manager [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Took 22.61 seconds to build instance. [ 1068.237717] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.244798] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.244798] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.244798] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore1] a5ea4075-616a-49ce-aa93-f1db495a022b {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.245095] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-947d539d-a8fe-440a-84fa-88f66d53486c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.255155] env[65107]: DEBUG oslo_vmware.api [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1068.255155] env[65107]: value = "task-5103442" [ 1068.255155] env[65107]: _type = "Task" [ 1068.255155] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.275678] env[65107]: DEBUG oslo_vmware.api [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103442, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.319739] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ef207f-0954-4292-9d8f-346c8ec42d3a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.332031] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b285e13-217e-4beb-92b4-003934ad4552 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.369269] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6577a02c-0a74-4e09-8f1d-1d5923d474a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.378966] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5037c511-4215-4d9f-9e44-1ba049595dc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.394937] env[65107]: DEBUG nova.compute.provider_tree [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.457598] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103439, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.475740] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52088bc8-8b78-70e6-1064-5eaad3052e0a, 'name': SearchDatastore_Task, 'duration_secs': 0.017483} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.477336] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90947397-ce2c-4f18-b324-2a0d8c27056b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.483789] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1068.483789] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be097a-2f99-ca61-f903-0802536b8d93" [ 1068.483789] env[65107]: _type = "Task" [ 1068.483789] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.494224] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be097a-2f99-ca61-f903-0802536b8d93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.719284] env[65107]: DEBUG oslo_concurrency.lockutils [req-c1469bb2-4984-4b81-82cc-d42d7f911094 req-086dfce4-9bd7-496b-8c94-db72ec65883b service nova] Releasing lock "refresh_cache-f6536493-fb80-4fb1-9ee1-e530734e9adf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.722193] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.733941] env[65107]: DEBUG oslo_concurrency.lockutils [None req-41c3c69f-a615-460a-b741-81ba7a6252bd tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.119s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.736019] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103441, 'name': PowerOffVM_Task, 'duration_secs': 0.296545} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.736019] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.736019] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.771398] env[65107]: DEBUG oslo_vmware.api [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103442, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237309} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.771670] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.771843] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.772033] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.772549] env[65107]: INFO nova.compute.manager [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1068.772549] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1068.772690] env[65107]: DEBUG nova.compute.manager [-] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1068.772751] env[65107]: DEBUG nova.network.neutron [-] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1068.772982] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.773590] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.773804] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.811662] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.898661] env[65107]: DEBUG nova.scheduler.client.report [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1068.961513] env[65107]: DEBUG oslo_vmware.api [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103439, 'name': PowerOnVM_Task, 'duration_secs': 0.796354} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.962821] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.964661] env[65107]: INFO nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Took 9.00 seconds to spawn the instance on the hypervisor. [ 1068.964661] env[65107]: DEBUG nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1068.964661] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7104c49d-61b5-45c0-86d6-174b8d221eb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.997196] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be097a-2f99-ca61-f903-0802536b8d93, 'name': SearchDatastore_Task, 'duration_secs': 0.037205} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.997581] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.997874] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] f6536493-fb80-4fb1-9ee1-e530734e9adf/f6536493-fb80-4fb1-9ee1-e530734e9adf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.998208] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c01f58b-f206-47f9-9887-8ff6a5af4dcf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.009043] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1069.009043] env[65107]: value = "task-5103443" [ 1069.009043] env[65107]: _type = "Task" [ 1069.009043] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.025810] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103443, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.247149] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1069.247149] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1069.247149] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1069.247472] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1069.247560] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1069.247700] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1069.247921] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.248086] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1069.248262] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1069.248436] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1069.248601] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1069.254340] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd01eb7c-8c62-4c32-917f-9dfa13b2d31e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.276556] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1069.276556] env[65107]: value = "task-5103444" [ 1069.276556] env[65107]: _type = "Task" [ 1069.276556] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.291922] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103444, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.413682] env[65107]: DEBUG nova.compute.manager [req-cf1276cd-2f12-46f8-bded-b0cd39065d39 req-d78d1b36-a9b2-4380-903c-bded3ba597f9 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Received event network-vif-deleted-2cf26308-e979-456c-92a8-f2bdf0d6948f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1069.413896] env[65107]: INFO nova.compute.manager [req-cf1276cd-2f12-46f8-bded-b0cd39065d39 req-d78d1b36-a9b2-4380-903c-bded3ba597f9 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Neutron deleted interface 2cf26308-e979-456c-92a8-f2bdf0d6948f; detaching it from the instance and deleting it from the info cache [ 1069.414086] env[65107]: DEBUG nova.network.neutron [req-cf1276cd-2f12-46f8-bded-b0cd39065d39 req-d78d1b36-a9b2-4380-903c-bded3ba597f9 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.486558] env[65107]: INFO nova.compute.manager [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Took 22.44 seconds to build instance. [ 1069.523500] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103443, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.546764] env[65107]: DEBUG nova.network.neutron [-] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.787227] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103444, 'name': ReconfigVM_Task, 'duration_secs': 0.220222} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.788063] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1069.917818] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.438s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.920841] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.199s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.922396] env[65107]: INFO nova.compute.claims [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.924735] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5145bc45-b8c9-4ded-80e0-26936babdcf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.934999] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4362ee3-ee33-497c-a41f-c85772c8b681 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.978665] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.979028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.981592] env[65107]: DEBUG nova.compute.manager [req-cf1276cd-2f12-46f8-bded-b0cd39065d39 req-d78d1b36-a9b2-4380-903c-bded3ba597f9 service nova] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Detach interface failed, port_id=2cf26308-e979-456c-92a8-f2bdf0d6948f, reason: Instance a5ea4075-616a-49ce-aa93-f1db495a022b could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1069.988633] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c2afa4f7-199a-40e6-b50f-0f56d861c329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.952s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.022285] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103443, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669141} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.022554] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] f6536493-fb80-4fb1-9ee1-e530734e9adf/f6536493-fb80-4fb1-9ee1-e530734e9adf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1070.022765] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1070.023556] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1de8fea-6c45-43f3-916d-00d783452338 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.031306] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1070.031306] env[65107]: value = "task-5103445" [ 1070.031306] env[65107]: _type = "Task" [ 1070.031306] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.041755] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103445, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.050518] env[65107]: INFO nova.compute.manager [-] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Took 1.28 seconds to deallocate network for instance. [ 1070.223967] env[65107]: DEBUG nova.compute.manager [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1070.232385] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "8dbed811-3d4e-4cae-9981-0334801013aa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.232629] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.232801] env[65107]: DEBUG nova.compute.manager [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1070.233696] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66830cc-4e37-4269-a628-11a4b5068d9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.241143] env[65107]: DEBUG nova.compute.manager [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1070.241847] env[65107]: DEBUG nova.objects.instance [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'flavor' on Instance uuid 8dbed811-3d4e-4cae-9981-0334801013aa {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.296832] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1070.297092] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1070.297383] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1070.297591] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1070.297739] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1070.297884] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1070.298909] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.299128] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1070.299310] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1070.299478] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1070.299654] env[65107]: DEBUG nova.virt.hardware [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1070.305437] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1070.306097] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83cb7c55-b982-4f2d-8d0b-45eb2e0d00a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.328020] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1070.328020] env[65107]: value = "task-5103446" [ 1070.328020] env[65107]: _type = "Task" [ 1070.328020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.335904] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103446, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.483433] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1070.487537] env[65107]: INFO nova.scheduler.client.report [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocation for migration 387dc95a-dd2c-4f87-8187-faf6d8f7234c [ 1070.543154] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103445, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.163595} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.543433] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1070.544262] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1dba1d-8ccc-46b6-8d52-562470da740a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.562343] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.571520] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] f6536493-fb80-4fb1-9ee1-e530734e9adf/f6536493-fb80-4fb1-9ee1-e530734e9adf.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.571846] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dae7eb3c-9ce6-4893-9779-ea62bb0cca00 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.591991] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1070.591991] env[65107]: value = "task-5103447" [ 1070.591991] env[65107]: _type = "Task" [ 1070.591991] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.601152] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103447, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.741511] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.839911] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103446, 'name': ReconfigVM_Task, 'duration_secs': 0.219562} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.840235] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1070.841197] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f84528a-746b-45c6-8b9b-2c30fa918e0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.867144] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.867557] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06e15ac5-a23f-45bd-abd8-a327f650859f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.887232] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1070.887232] env[65107]: value = "task-5103448" [ 1070.887232] env[65107]: _type = "Task" [ 1070.887232] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.900326] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103448, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.996191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b32bcec-de23-483c-93c7-a427ef2e51af tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.681s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.013146] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.103652] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103447, 'name': ReconfigVM_Task, 'duration_secs': 0.278256} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.106910] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Reconfigured VM instance instance-00000063 to attach disk [datastore1] f6536493-fb80-4fb1-9ee1-e530734e9adf/f6536493-fb80-4fb1-9ee1-e530734e9adf.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.107939] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7c80230-5fa5-4470-8090-993c944d1386 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.115809] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1071.115809] env[65107]: value = "task-5103449" [ 1071.115809] env[65107]: _type = "Task" [ 1071.115809] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.128289] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103449, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.212935] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a4b625-f2bb-4d46-9683-b818f886cc71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.221524] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69a292b-8d0e-44d6-8376-d8a632f5b5f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.258182] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.258819] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-551e3cbf-a209-49d1-9f84-47ad5547cd46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.261266] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8be6dd-eef7-4bd7-babd-42e06f234508 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.271347] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d5e857-e25c-402a-91c6-17c204a90f57 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.275850] env[65107]: DEBUG oslo_vmware.api [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1071.275850] env[65107]: value = "task-5103450" [ 1071.275850] env[65107]: _type = "Task" [ 1071.275850] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.288568] env[65107]: DEBUG nova.compute.provider_tree [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.295960] env[65107]: DEBUG oslo_vmware.api [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.398400] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103448, 'name': ReconfigVM_Task, 'duration_secs': 0.511049} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.398837] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.399024] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1071.625928] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103449, 'name': Rename_Task, 'duration_secs': 0.176504} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.626246] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.626511] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea9a47da-c4c3-40d4-8093-a6420509eeb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.633837] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1071.633837] env[65107]: value = "task-5103451" [ 1071.633837] env[65107]: _type = "Task" [ 1071.633837] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.646733] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.786050] env[65107]: DEBUG oslo_vmware.api [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103450, 'name': PowerOffVM_Task, 'duration_secs': 0.255403} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.786371] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.786575] env[65107]: DEBUG nova.compute.manager [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1071.787398] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf08c5c9-7e4f-4f14-bfb8-478be0b8caf8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.792437] env[65107]: DEBUG nova.scheduler.client.report [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1071.906059] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a941bdf7-5489-47b3-a9f6-b3522a4c19bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.930600] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fbbec6-ff93-4d3f-bdec-53449467a69a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.953661] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1072.146719] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103451, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.300418] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.300988] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1072.306044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.744s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.306351] env[65107]: DEBUG nova.objects.instance [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'resources' on Instance uuid a5ea4075-616a-49ce-aa93-f1db495a022b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.307487] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6359b4a-5f52-4f23-9575-b28ed9deec32 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.075s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.356647] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.356915] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.357165] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.357477] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.357650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.360343] env[65107]: INFO nova.compute.manager [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Terminating instance [ 1072.645722] env[65107]: DEBUG oslo_vmware.api [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103451, 'name': PowerOnVM_Task, 'duration_secs': 0.754709} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.645949] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.646203] env[65107]: INFO nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Took 8.32 seconds to spawn the instance on the hypervisor. [ 1072.646394] env[65107]: DEBUG nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1072.647203] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c131fcba-6d64-4fcc-aa0c-827fb2edfcb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.807908] env[65107]: DEBUG nova.compute.utils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1072.809548] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1072.810162] env[65107]: DEBUG nova.network.neutron [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1072.810548] env[65107]: WARNING neutronclient.v2_0.client [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.811020] env[65107]: WARNING neutronclient.v2_0.client [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.812306] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.812306] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.866695] env[65107]: DEBUG nova.compute.manager [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1072.866909] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.868395] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e834f09b-717b-4a94-8e0b-64affcd41ebf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.874891] env[65107]: DEBUG nova.policy [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b03a2312b7d643dd85fceac3d0226ea1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d2bec7b22c74eaa8cbc9c85ebea1e68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1072.881910] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.885544] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8331f108-86be-4c9a-b03b-f533cf78d346 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.891157] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "3f544b3f-703a-49d5-b3e8-16672766f691" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.891383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.896201] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1072.896201] env[65107]: value = "task-5103452" [ 1072.896201] env[65107]: _type = "Task" [ 1072.896201] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.909364] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.928662] env[65107]: INFO nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Rebuilding instance [ 1072.980597] env[65107]: DEBUG nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1072.981485] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579c5876-ea54-4171-8c69-6327559e1fb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.118776] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8734e35d-ebd6-43b8-9c5b-09671a9ae32b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.129254] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7775cba6-112e-493f-925f-6e4507f6b68d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.176522] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0690e68a-0421-4147-91ea-e753865b8e80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.182723] env[65107]: INFO nova.compute.manager [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Took 16.25 seconds to build instance. [ 1073.188638] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75fbccf-12d1-49cb-8302-fb57cd8a255d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.206122] env[65107]: DEBUG nova.compute.provider_tree [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.226627] env[65107]: DEBUG nova.network.neutron [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Successfully created port: 5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1073.322975] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1073.394389] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1073.410653] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103452, 'name': PowerOffVM_Task, 'duration_secs': 0.243321} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.410995] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.412270] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.412270] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d752f844-6d5b-4cc2-9a35-577ab845d230 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.493350] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.493607] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.493789] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore2] 00574b77-dad6-4f0a-bbcc-20a2b4a5df65 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.495551] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-396c4d5a-dac8-4edd-a8b4-64fb3f8b1b4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.506372] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1073.506372] env[65107]: value = "task-5103454" [ 1073.506372] env[65107]: _type = "Task" [ 1073.506372] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.518021] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.582627] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.582964] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.672170] env[65107]: DEBUG nova.network.neutron [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Port 416a2590-6a42-4a32-a922-a9d5f95786d6 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1073.684482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-bb410d19-0589-4967-b253-01b5a81f8a2e tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.759s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.709331] env[65107]: DEBUG nova.scheduler.client.report [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1073.917844] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.972130] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "f6536493-fb80-4fb1-9ee1-e530734e9adf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.972406] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1073.972614] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "f6536493-fb80-4fb1-9ee1-e530734e9adf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.972794] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1073.972961] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.975607] env[65107]: INFO nova.compute.manager [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Terminating instance [ 1073.999852] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.000138] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-485c6b24-a0ae-4cd8-a8d9-12b3ebb21708 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.007670] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1074.007670] env[65107]: value = "task-5103455" [ 1074.007670] env[65107]: _type = "Task" [ 1074.007670] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.020018] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.023561] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1074.023798] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1074.024640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab90d497-cd29-4d71-9407-99bc7ef1fb4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.031424] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.031645] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0834f0e-efd3-4474-b83b-4c3866d84fac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.110100] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.110342] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.110532] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore1] 8dbed811-3d4e-4cae-9981-0334801013aa {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.110860] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6579de6d-4304-400e-bdb2-2b0712d84dbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.118211] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1074.118211] env[65107]: value = "task-5103457" [ 1074.118211] env[65107]: _type = "Task" [ 1074.118211] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.128848] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.214333] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.216907] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.475s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.237377] env[65107]: INFO nova.scheduler.client.report [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance a5ea4075-616a-49ce-aa93-f1db495a022b [ 1074.333944] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1074.361714] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1074.362071] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1074.362241] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1074.362489] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1074.362679] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1074.362867] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1074.363152] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.363369] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1074.363582] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1074.363789] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1074.364014] env[65107]: DEBUG nova.virt.hardware [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1074.365017] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5f1d5a-db48-4334-85f6-a50e96bdf4ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.373533] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b088bc-3cd4-4c79-9346-83c22b18992b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.479581] env[65107]: DEBUG nova.compute.manager [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1074.479819] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1074.480742] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01261147-4791-4927-9f5f-e69cdc86fbdb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.488468] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.488700] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-379c0245-28a5-4ca7-b89f-48909d7e7509 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.496204] env[65107]: DEBUG oslo_vmware.api [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1074.496204] env[65107]: value = "task-5103458" [ 1074.496204] env[65107]: _type = "Task" [ 1074.496204] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.504971] env[65107]: DEBUG oslo_vmware.api [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103458, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.518547] env[65107]: DEBUG oslo_vmware.api [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.656818} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.518761] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.518953] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.519153] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.519334] env[65107]: INFO nova.compute.manager [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1074.519584] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1074.519792] env[65107]: DEBUG nova.compute.manager [-] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1074.519885] env[65107]: DEBUG nova.network.neutron [-] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1074.520155] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.520746] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.521022] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.561052] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.629572] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227958} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.629847] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.630071] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.630263] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.694914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.695163] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.695372] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.723428] env[65107]: INFO nova.compute.claims [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.745229] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5287b9cf-1367-45e2-b1d9-75f2b05ebee9 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "a5ea4075-616a-49ce-aa93-f1db495a022b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.629s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.819173] env[65107]: DEBUG nova.network.neutron [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Successfully updated port: 5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1075.009248] env[65107]: DEBUG oslo_vmware.api [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103458, 'name': PowerOffVM_Task, 'duration_secs': 0.21387} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.009536] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.009736] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1075.009954] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15bf3e47-38f0-4b66-9e70-d83956839746 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.021283] env[65107]: DEBUG nova.compute.manager [req-5e5ea113-bf51-4151-a3ff-afece0219232 req-561cfd92-772d-42cc-895c-6016129f2a13 service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Received event network-vif-plugged-5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1075.021642] env[65107]: DEBUG oslo_concurrency.lockutils [req-5e5ea113-bf51-4151-a3ff-afece0219232 req-561cfd92-772d-42cc-895c-6016129f2a13 service nova] Acquiring lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.021714] env[65107]: DEBUG oslo_concurrency.lockutils [req-5e5ea113-bf51-4151-a3ff-afece0219232 req-561cfd92-772d-42cc-895c-6016129f2a13 service nova] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.021861] env[65107]: DEBUG oslo_concurrency.lockutils [req-5e5ea113-bf51-4151-a3ff-afece0219232 req-561cfd92-772d-42cc-895c-6016129f2a13 service nova] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.022034] env[65107]: DEBUG nova.compute.manager [req-5e5ea113-bf51-4151-a3ff-afece0219232 req-561cfd92-772d-42cc-895c-6016129f2a13 service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] No waiting events found dispatching network-vif-plugged-5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1075.022210] env[65107]: WARNING nova.compute.manager [req-5e5ea113-bf51-4151-a3ff-afece0219232 req-561cfd92-772d-42cc-895c-6016129f2a13 service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Received unexpected event network-vif-plugged-5343cf6f-59a3-4f9c-a534-449b9816168a for instance with vm_state building and task_state spawning. [ 1075.072581] env[65107]: DEBUG nova.compute.manager [req-4365a8a5-7d10-4644-9f64-0e2c5bd0edbd req-eb3ebf49-d6fa-4bbd-a416-8c4e773860ca service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Received event network-vif-deleted-d88e66c9-ba7c-452d-8e44-8c9603f2b258 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1075.073570] env[65107]: INFO nova.compute.manager [req-4365a8a5-7d10-4644-9f64-0e2c5bd0edbd req-eb3ebf49-d6fa-4bbd-a416-8c4e773860ca service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Neutron deleted interface d88e66c9-ba7c-452d-8e44-8c9603f2b258; detaching it from the instance and deleting it from the info cache [ 1075.073779] env[65107]: DEBUG nova.network.neutron [req-4365a8a5-7d10-4644-9f64-0e2c5bd0edbd req-eb3ebf49-d6fa-4bbd-a416-8c4e773860ca service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1075.081041] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1075.081239] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1075.081426] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Deleting the datastore file [datastore1] f6536493-fb80-4fb1-9ee1-e530734e9adf {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.081701] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9640425-720b-4835-9d49-a7660886f02c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.091037] env[65107]: DEBUG oslo_vmware.api [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for the task: (returnval){ [ 1075.091037] env[65107]: value = "task-5103460" [ 1075.091037] env[65107]: _type = "Task" [ 1075.091037] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.103255] env[65107]: DEBUG oslo_vmware.api [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103460, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.231571] env[65107]: INFO nova.compute.resource_tracker [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating resource usage from migration a9997a68-5b5b-4571-b870-f31a9bee6557 [ 1075.276116] env[65107]: DEBUG nova.network.neutron [-] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1075.322309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.322496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.322759] env[65107]: DEBUG nova.network.neutron [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1075.483248] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e85912-c3af-4728-a173-260efd34535c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.491867] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64ac936-882a-4b0d-ac66-3dc19a1a8bfd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.525151] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a275ba4c-c7d1-4532-891d-179f4240b168 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.533526] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85283267-16f5-448d-bb48-1f6bcea3d26b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.548507] env[65107]: DEBUG nova.compute.provider_tree [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.576651] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02d83ab4-9f52-41cd-b3c9-7b74bf65cb8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.587725] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749463e5-a367-42b1-ace7-a663e37c8d3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.613008] env[65107]: DEBUG oslo_vmware.api [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Task: {'id': task-5103460, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311608} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.613233] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.613383] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.613644] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.613925] env[65107]: INFO nova.compute.manager [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1075.614347] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1075.614486] env[65107]: DEBUG nova.compute.manager [-] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1075.614543] env[65107]: DEBUG nova.network.neutron [-] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1075.614798] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.615380] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.615643] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.636320] env[65107]: DEBUG nova.compute.manager [req-4365a8a5-7d10-4644-9f64-0e2c5bd0edbd req-eb3ebf49-d6fa-4bbd-a416-8c4e773860ca service nova] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Detach interface failed, port_id=d88e66c9-ba7c-452d-8e44-8c9603f2b258, reason: Instance 00574b77-dad6-4f0a-bbcc-20a2b4a5df65 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1075.673797] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1075.674024] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1075.674161] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1075.674347] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1075.674492] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1075.674639] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1075.674850] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.675016] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1075.675700] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1075.676264] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1075.676264] env[65107]: DEBUG nova.virt.hardware [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1075.677281] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56df9f6-a8e4-48f0-bbab-7e95ceb2089c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.684901] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.687321] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.687576] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.694027] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc732bd-2f91-4124-9dc3-291f43f14ff4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.702044] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.715352] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:c8:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6091d1cb-29b9-4a22-8827-990e8ea7d318', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.722515] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1075.723085] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1075.723350] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-098d5f68-f510-4d49-b976-63ca27d0172c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.745177] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.745177] env[65107]: value = "task-5103461" [ 1075.745177] env[65107]: _type = "Task" [ 1075.745177] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.746468] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.746652] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.746829] env[65107]: DEBUG nova.network.neutron [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1075.757498] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103461, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.779714] env[65107]: INFO nova.compute.manager [-] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Took 1.26 seconds to deallocate network for instance. [ 1075.824967] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.825484] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.862331] env[65107]: DEBUG nova.network.neutron [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1075.885565] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.886936] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.934449] env[65107]: WARNING neutronclient.v2_0.client [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.935201] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.935627] env[65107]: WARNING openstack [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.020742] env[65107]: DEBUG nova.network.neutron [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Updating instance_info_cache with network_info: [{"id": "5343cf6f-59a3-4f9c-a534-449b9816168a", "address": "fa:16:3e:db:32:b1", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5343cf6f-59", "ovs_interfaceid": "5343cf6f-59a3-4f9c-a534-449b9816168a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1076.053045] env[65107]: DEBUG nova.scheduler.client.report [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1076.190705] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1076.252287] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.252828] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.253202] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.267797] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103461, 'name': CreateVM_Task, 'duration_secs': 0.381275} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.268018] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1076.268484] env[65107]: WARNING neutronclient.v2_0.client [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1076.268854] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.269012] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.269345] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1076.270511] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2cbf883-08a2-4293-97d9-e174d764f697 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.275883] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1076.275883] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523794c6-ec5e-d8ab-d054-0876de9cec12" [ 1076.275883] env[65107]: _type = "Task" [ 1076.275883] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.284689] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523794c6-ec5e-d8ab-d054-0876de9cec12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.288295] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.384954] env[65107]: DEBUG nova.network.neutron [-] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1076.387817] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.388445] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.459852] env[65107]: WARNING neutronclient.v2_0.client [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.460552] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.460907] env[65107]: WARNING openstack [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.523558] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.523872] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Instance network_info: |[{"id": "5343cf6f-59a3-4f9c-a534-449b9816168a", "address": "fa:16:3e:db:32:b1", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5343cf6f-59", "ovs_interfaceid": "5343cf6f-59a3-4f9c-a534-449b9816168a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1076.524421] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:32:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5343cf6f-59a3-4f9c-a534-449b9816168a', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.532133] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Creating folder: Project (3d2bec7b22c74eaa8cbc9c85ebea1e68). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1076.532457] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15b02331-6c34-4e14-a27a-48a13015cbe8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.544674] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Created folder: Project (3d2bec7b22c74eaa8cbc9c85ebea1e68) in parent group-v992574. [ 1076.544902] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Creating folder: Instances. Parent ref: group-v992856. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1076.547485] env[65107]: DEBUG nova.network.neutron [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1076.548794] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-850d4355-b8c1-442d-9af2-6a13ab538d55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.557682] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.341s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.557887] env[65107]: INFO nova.compute.manager [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Migrating [ 1076.567308] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.556s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.568843] env[65107]: INFO nova.compute.claims [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.572924] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Created folder: Instances in parent group-v992856. [ 1076.572924] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1076.575635] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.580520] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b40652b3-7bc9-4455-9ce0-0b781fa57530 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.601536] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.601637] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.601813] env[65107]: DEBUG nova.network.neutron [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1076.609198] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.609198] env[65107]: value = "task-5103464" [ 1076.609198] env[65107]: _type = "Task" [ 1076.609198] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.621605] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103464, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.717562] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.786876] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523794c6-ec5e-d8ab-d054-0876de9cec12, 'name': SearchDatastore_Task, 'duration_secs': 0.010251} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.787288] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.787568] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1076.787811] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.787958] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.788154] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1076.788427] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-060592f7-39c4-4954-a4d7-55c8ae0e5102 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.797562] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1076.797733] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1076.798474] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e639559d-3a15-4ff4-8ca5-abaefd31c9c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.804600] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1076.804600] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be7967-257f-b5f6-a229-3f0aaf7b9437" [ 1076.804600] env[65107]: _type = "Task" [ 1076.804600] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.812638] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be7967-257f-b5f6-a229-3f0aaf7b9437, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.895969] env[65107]: INFO nova.compute.manager [-] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Took 1.28 seconds to deallocate network for instance. [ 1077.051911] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.062031] env[65107]: DEBUG nova.compute.manager [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Received event network-changed-5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1077.062470] env[65107]: DEBUG nova.compute.manager [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Refreshing instance network info cache due to event network-changed-5343cf6f-59a3-4f9c-a534-449b9816168a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1077.062777] env[65107]: DEBUG oslo_concurrency.lockutils [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Acquiring lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.062955] env[65107]: DEBUG oslo_concurrency.lockutils [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Acquired lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.063198] env[65107]: DEBUG nova.network.neutron [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Refreshing network info cache for port 5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1077.104392] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.105177] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.105542] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.124524] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103464, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.214814] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.215245] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.281755] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.282465] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.282829] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.316659] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52be7967-257f-b5f6-a229-3f0aaf7b9437, 'name': SearchDatastore_Task, 'duration_secs': 0.010383} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.317773] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a514386b-938e-43a5-be69-89fbd74f6eab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.325233] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1077.325233] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5276030a-23bb-f3cb-0456-60de464c80de" [ 1077.325233] env[65107]: _type = "Task" [ 1077.325233] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.334953] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5276030a-23bb-f3cb-0456-60de464c80de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.377596] env[65107]: DEBUG nova.network.neutron [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [{"id": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "address": "fa:16:3e:44:a8:f3", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e9a520d-89", "ovs_interfaceid": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.403611] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1077.563630] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32b5e98-3b06-42a7-a785-82bdac1373fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.566687] env[65107]: WARNING neutronclient.v2_0.client [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.567370] env[65107]: WARNING openstack [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.567716] env[65107]: WARNING openstack [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.583099] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868728b7-28d4-49b5-ab1b-3ac712ff9a4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.630752] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103464, 'name': CreateVM_Task, 'duration_secs': 0.523853} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.631213] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.631689] env[65107]: WARNING neutronclient.v2_0.client [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.631957] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.632131] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.632455] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1077.632720] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d71630f6-37ed-4255-8696-cb2769a2bfa6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.642322] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1077.642322] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ddddcd-0345-8a5a-4ccb-a068ebc08bd8" [ 1077.642322] env[65107]: _type = "Task" [ 1077.642322] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.659470] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ddddcd-0345-8a5a-4ccb-a068ebc08bd8, 'name': SearchDatastore_Task, 'duration_secs': 0.011019} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.659470] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.659577] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.659735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.669571] env[65107]: WARNING openstack [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.670049] env[65107]: WARNING openstack [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.723675] env[65107]: WARNING neutronclient.v2_0.client [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.724356] env[65107]: WARNING openstack [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.724697] env[65107]: WARNING openstack [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.804999] env[65107]: DEBUG nova.network.neutron [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Updated VIF entry in instance network info cache for port 5343cf6f-59a3-4f9c-a534-449b9816168a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1077.805387] env[65107]: DEBUG nova.network.neutron [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Updating instance_info_cache with network_info: [{"id": "5343cf6f-59a3-4f9c-a534-449b9816168a", "address": "fa:16:3e:db:32:b1", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5343cf6f-59", "ovs_interfaceid": "5343cf6f-59a3-4f9c-a534-449b9816168a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.835848] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5276030a-23bb-f3cb-0456-60de464c80de, 'name': SearchDatastore_Task, 'duration_secs': 0.022707} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.838482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.838744] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1077.839225] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.839418] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.839635] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2fb40632-7b05-4a95-b74a-527c2b947913 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.841630] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a4cab77-5a58-43b4-8ba1-ff582784f3bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.851359] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1077.851359] env[65107]: value = "task-5103465" [ 1077.851359] env[65107]: _type = "Task" [ 1077.851359] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.852789] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.853065] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.856571] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b6ac595-2b1c-4695-8e03-0c46d2918efd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.861288] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ed50d2-c7cc-4ce7-9ffe-cd19b88ee55b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.868684] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1077.868684] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520ae085-af25-1c08-d2b8-7eaa14068948" [ 1077.868684] env[65107]: _type = "Task" [ 1077.868684] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.868976] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.875901] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7fef6aa-2636-4abf-939f-789cc6e7271a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.882772] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.884240] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520ae085-af25-1c08-d2b8-7eaa14068948, 'name': SearchDatastore_Task, 'duration_secs': 0.009713} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.886450] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09c5fa43-5ba1-4c06-87db-0f104b4dbe21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.916304] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc694f7a-02d1-4471-9aa3-c34a0bab734a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.920881] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1077.920881] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527390bf-4dfc-8863-6393-7493b370b149" [ 1077.920881] env[65107]: _type = "Task" [ 1077.920881] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.930665] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d3ccf5-710c-4040-834f-cc6a53cf41fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.938961] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527390bf-4dfc-8863-6393-7493b370b149, 'name': SearchDatastore_Task, 'duration_secs': 0.008827} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.939651] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.939952] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1077.940239] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86b53eb9-1147-4630-becb-5b14dad3675c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.950918] env[65107]: DEBUG nova.compute.provider_tree [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.957930] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1077.957930] env[65107]: value = "task-5103466" [ 1077.957930] env[65107]: _type = "Task" [ 1077.957930] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.967094] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103466, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.309022] env[65107]: DEBUG oslo_concurrency.lockutils [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] Releasing lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.309022] env[65107]: DEBUG nova.compute.manager [req-f6bcd0c3-f58e-4919-a38e-287b9b83b8d5 req-942b3215-b885-4680-9e51-90963af88c0f service nova] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Received event network-vif-deleted-32156605-37d7-4888-9a96-3ac243961051 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1078.362977] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476368} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.363164] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.363309] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.363529] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2230628b-3a08-469d-825f-5a54eba936e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.371080] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1078.371080] env[65107]: value = "task-5103467" [ 1078.371080] env[65107]: _type = "Task" [ 1078.371080] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.381652] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.454298] env[65107]: DEBUG nova.scheduler.client.report [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.472543] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103466, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.706166] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0357f2f2-31fb-4a2b-b4b7-055aa5c44d2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.728164] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b24b4d0-fab3-4c1a-a545-bc5542c2f4ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.736328] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1078.883520] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083352} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.883814] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1078.884704] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d712cd8d-7f2e-4e1e-9ca7-5d51e69d7710 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.910052] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.911135] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d2fba60-4b22-4477-a1e9-4195820ff073 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.931538] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1078.931538] env[65107]: value = "task-5103468" [ 1078.931538] env[65107]: _type = "Task" [ 1078.931538] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.940812] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103468, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.960221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.960725] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1078.963490] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.046s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1078.964907] env[65107]: INFO nova.compute.claims [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.977364] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103466, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.716549} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.977608] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.977819] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.978114] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63381bc2-0857-4489-a770-57aa875d6c72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.984981] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1078.984981] env[65107]: value = "task-5103469" [ 1078.984981] env[65107]: _type = "Task" [ 1078.984981] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.995235] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.242758] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1079.243095] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c15ddd99-bfe5-46d5-a01f-654adf8ddeca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.252967] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1079.252967] env[65107]: value = "task-5103470" [ 1079.252967] env[65107]: _type = "Task" [ 1079.252967] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.262821] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103470, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.415017] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05600e2d-6345-41ad-ac96-d6fcd99c46fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.434373] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.446982] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.476704] env[65107]: DEBUG nova.compute.utils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1079.478358] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1079.478579] env[65107]: DEBUG nova.network.neutron [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1079.478930] env[65107]: WARNING neutronclient.v2_0.client [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.479278] env[65107]: WARNING neutronclient.v2_0.client [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.479909] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.480306] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.500155] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060633} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.500155] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.500859] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a670ab-56cc-47ab-b49b-850896a98d2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.526345] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.530229] env[65107]: DEBUG nova.policy [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1079.532209] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73ca04b1-0c59-4e9d-a279-495adf5565a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.555097] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1079.555097] env[65107]: value = "task-5103471" [ 1079.555097] env[65107]: _type = "Task" [ 1079.555097] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.564744] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103471, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.765014] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103470, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.879391] env[65107]: DEBUG nova.network.neutron [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Successfully created port: 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1079.944689] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.945448] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103468, 'name': ReconfigVM_Task, 'duration_secs': 0.927183} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.945692] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-490ed922-f965-493f-9952-423b54ee6a66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.947528] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 8dbed811-3d4e-4cae-9981-0334801013aa/8dbed811-3d4e-4cae-9981-0334801013aa.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.948221] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c45943b-92d2-44c7-b7f1-21a2d7250d71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.955361] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1079.955361] env[65107]: value = "task-5103473" [ 1079.955361] env[65107]: _type = "Task" [ 1079.955361] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.956831] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1079.956831] env[65107]: value = "task-5103472" [ 1079.956831] env[65107]: _type = "Task" [ 1079.956831] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.968825] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103473, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.971973] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.989503] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1080.070771] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103471, 'name': ReconfigVM_Task, 'duration_secs': 0.378843} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.071120] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.071855] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b3d46f3-852e-42b4-b3a2-b30d83f40165 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.083029] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1080.083029] env[65107]: value = "task-5103474" [ 1080.083029] env[65107]: _type = "Task" [ 1080.083029] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.089520] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103474, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.270382] env[65107]: DEBUG oslo_vmware.api [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103470, 'name': PowerOnVM_Task, 'duration_secs': 0.655313} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.274197] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1080.274491] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe0ff04-2c1e-4443-8f87-3531f76eb0d7 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance '8cddc4d1-c94c-4f2b-9c4b-3125e4638050' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1080.322016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26638d1b-eade-4b35-9bc0-cba68d80b1cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.330756] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74491003-3b63-4ab0-bcc4-e9078eb5ddd5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.367072] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1d48ea-f419-43e0-b56b-961c31e23ab9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.380159] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81efeac0-4363-4e40-b291-99675b964f92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.396856] env[65107]: DEBUG nova.compute.provider_tree [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.474179] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103473, 'name': Rename_Task, 'duration_secs': 0.173177} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.475058] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103472, 'name': PowerOffVM_Task, 'duration_secs': 0.211146} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.475108] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.476989] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.476989] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1080.478990] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24a4c7f4-e310-46dd-8707-bfec966d712f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.486486] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1080.486486] env[65107]: value = "task-5103475" [ 1080.486486] env[65107]: _type = "Task" [ 1080.486486] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.499770] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.592706] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103474, 'name': Rename_Task, 'duration_secs': 0.163503} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.592995] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.593304] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88c2a1a2-3d77-4858-90fb-ec8d0e39cae4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.600298] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1080.600298] env[65107]: value = "task-5103476" [ 1080.600298] env[65107]: _type = "Task" [ 1080.600298] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.609842] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103476, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.901522] env[65107]: DEBUG nova.scheduler.client.report [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1080.984446] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1080.984695] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1080.984854] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1080.985042] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1080.985265] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1080.985428] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1080.985725] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.985961] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1080.986181] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1080.986357] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1080.986526] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1080.992097] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65b0af07-8574-4ad9-907d-c548995415c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.008168] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1081.019096] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103475, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.021415] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1081.021415] env[65107]: value = "task-5103477" [ 1081.021415] env[65107]: _type = "Task" [ 1081.021415] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.034269] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103477, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.045625] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1081.045852] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1081.045999] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1081.046240] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1081.046428] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1081.046596] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1081.046813] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.046972] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1081.047163] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1081.047332] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1081.047508] env[65107]: DEBUG nova.virt.hardware [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1081.048464] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94572d83-7e07-4dff-a090-993601cd85c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.064963] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a95c224-014a-4da7-a867-0fdcec01bcc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.112687] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103476, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.408906] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.409484] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1081.412223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.124s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.412418] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.414624] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.697s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.416257] env[65107]: INFO nova.compute.claims [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1081.440321] env[65107]: INFO nova.scheduler.client.report [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocations for instance 00574b77-dad6-4f0a-bbcc-20a2b4a5df65 [ 1081.502078] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103475, 'name': PowerOnVM_Task, 'duration_secs': 0.598257} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.502431] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.502577] env[65107]: DEBUG nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1081.503425] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28120a06-77a3-49cb-8503-3f28ee5530ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.531306] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103477, 'name': ReconfigVM_Task, 'duration_secs': 0.272128} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.531631] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1081.612555] env[65107]: DEBUG oslo_vmware.api [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103476, 'name': PowerOnVM_Task, 'duration_secs': 0.535789} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.612841] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.613054] env[65107]: INFO nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Took 7.28 seconds to spawn the instance on the hypervisor. [ 1081.613273] env[65107]: DEBUG nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1081.614100] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e994e6f0-b264-4cc8-938c-b5cba4357050 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.804783] env[65107]: DEBUG nova.compute.manager [req-81f1e8f1-8f07-41ef-9731-dd278f798dfe req-f106650b-1f42-4ee3-a975-cd66c58dbd7d service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-vif-plugged-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1081.805025] env[65107]: DEBUG oslo_concurrency.lockutils [req-81f1e8f1-8f07-41ef-9731-dd278f798dfe req-f106650b-1f42-4ee3-a975-cd66c58dbd7d service nova] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.805250] env[65107]: DEBUG oslo_concurrency.lockutils [req-81f1e8f1-8f07-41ef-9731-dd278f798dfe req-f106650b-1f42-4ee3-a975-cd66c58dbd7d service nova] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.805420] env[65107]: DEBUG oslo_concurrency.lockutils [req-81f1e8f1-8f07-41ef-9731-dd278f798dfe req-f106650b-1f42-4ee3-a975-cd66c58dbd7d service nova] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.805591] env[65107]: DEBUG nova.compute.manager [req-81f1e8f1-8f07-41ef-9731-dd278f798dfe req-f106650b-1f42-4ee3-a975-cd66c58dbd7d service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] No waiting events found dispatching network-vif-plugged-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1081.805757] env[65107]: WARNING nova.compute.manager [req-81f1e8f1-8f07-41ef-9731-dd278f798dfe req-f106650b-1f42-4ee3-a975-cd66c58dbd7d service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received unexpected event network-vif-plugged-47322e57-eaa6-4bb8-98ec-3098749bd52c for instance with vm_state building and task_state spawning. [ 1081.889210] env[65107]: DEBUG nova.network.neutron [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Successfully updated port: 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1081.921758] env[65107]: DEBUG nova.compute.utils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1081.925262] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1081.925935] env[65107]: DEBUG nova.network.neutron [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1081.925935] env[65107]: WARNING neutronclient.v2_0.client [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.926546] env[65107]: WARNING neutronclient.v2_0.client [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.926811] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.927071] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.949644] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7351f651-3a4d-44d4-9de6-daba2134d629 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "00574b77-dad6-4f0a-bbcc-20a2b4a5df65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.593s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.981259] env[65107]: DEBUG nova.policy [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '365eed31877241e58fc860ed5620d5f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c805fe68d0fd4147b9b4b7fdc9d6a6d0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1082.014812] env[65107]: INFO nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] bringing vm to original state: 'stopped' [ 1082.038890] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1082.039917] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1082.039917] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1082.039917] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1082.040255] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1082.040380] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1082.040679] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.040857] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1082.041048] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1082.041220] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1082.041397] env[65107]: DEBUG nova.virt.hardware [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1082.047079] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1082.047408] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1224fa79-367d-4ba0-a31a-ed161146300c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.068508] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1082.068508] env[65107]: value = "task-5103478" [ 1082.068508] env[65107]: _type = "Task" [ 1082.068508] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.079128] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103478, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.134057] env[65107]: INFO nova.compute.manager [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Took 13.43 seconds to build instance. [ 1082.340897] env[65107]: DEBUG nova.network.neutron [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Successfully created port: 38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1082.392421] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.392645] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.392786] env[65107]: DEBUG nova.network.neutron [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1082.425951] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1082.583546] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103478, 'name': ReconfigVM_Task, 'duration_secs': 0.181934} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.583897] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1082.584741] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccaa71d8-1200-4d8f-b285-f921150d3903 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.613575] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b/7373fbbe-52c1-4955-8da9-a5563d71391b.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.617035] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c587379-0597-47b6-a442-6553ba47777b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.634982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e64c7e57-2b7d-4b29-9529-9ef28cdf8dd3 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.946s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.638184] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1082.638184] env[65107]: value = "task-5103479" [ 1082.638184] env[65107]: _type = "Task" [ 1082.638184] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.648954] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103479, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.750218] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88197fe3-9b5d-4935-8925-9147ef618890 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.757988] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd58309-79ae-4bf1-91cf-9ce9d8ed8390 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.794998] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee00207-c2b5-4b08-851d-71dc571500c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.803920] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c77cf43-92d5-4fd1-a455-09dd17e288b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.819946] env[65107]: DEBUG nova.compute.provider_tree [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.896315] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.899412] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.899475] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.962299] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.962707] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.962971] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.992502] env[65107]: DEBUG nova.network.neutron [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1083.021698] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "8dbed811-3d4e-4cae-9981-0334801013aa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.022100] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.022357] env[65107]: DEBUG nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1083.024227] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4dfd9a-e0e1-4459-a129-f5cab9231de6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.033677] env[65107]: DEBUG nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1083.049462] env[65107]: DEBUG nova.network.neutron [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Port 416a2590-6a42-4a32-a922-a9d5f95786d6 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1083.049876] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.050138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.050364] env[65107]: DEBUG nova.network.neutron [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1083.065558] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.066155] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.149839] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103479, 'name': ReconfigVM_Task, 'duration_secs': 0.300411} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.150356] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b/7373fbbe-52c1-4955-8da9-a5563d71391b.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.150731] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.230863] env[65107]: WARNING neutronclient.v2_0.client [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.231640] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.232047] env[65107]: WARNING openstack [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.323384] env[65107]: DEBUG nova.scheduler.client.report [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1083.336257] env[65107]: INFO nova.compute.manager [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Rescuing [ 1083.336533] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.336674] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.336883] env[65107]: DEBUG nova.network.neutron [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1083.382356] env[65107]: DEBUG nova.network.neutron [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1083.437310] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1083.467081] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1083.467382] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1083.467535] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1083.467709] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1083.467848] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1083.467989] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1083.468282] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1083.468458] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1083.468625] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1083.468783] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1083.468954] env[65107]: DEBUG nova.virt.hardware [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1083.470211] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33d74ad-1d26-4e71-9521-ec3661b17c5e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.479410] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e83482-481e-4001-a082-dc7e9e7bb5bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.541761] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.542040] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ee170cd-def3-4c20-abd9-e835ea0b8b48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.550690] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1083.550690] env[65107]: value = "task-5103480" [ 1083.550690] env[65107]: _type = "Task" [ 1083.550690] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.556120] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.556893] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.557230] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.570658] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.658596] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d45585-a4c2-4131-9bfd-97b8da891fac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.684828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b86243-b3d4-46fb-8a9a-b69883bc8f87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.704267] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.732652] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.733327] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.797427] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.798093] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.798513] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.830638] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.831282] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1083.834503] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.431s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.834771] env[65107]: DEBUG nova.objects.instance [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lazy-loading 'resources' on Instance uuid f6536493-fb80-4fb1-9ee1-e530734e9adf {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.842021] env[65107]: WARNING neutronclient.v2_0.client [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.842831] env[65107]: WARNING openstack [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.843304] env[65107]: WARNING openstack [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.858036] env[65107]: DEBUG nova.compute.manager [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1083.858238] env[65107]: DEBUG nova.compute.manager [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing instance network info cache due to event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1083.858462] env[65107]: DEBUG oslo_concurrency.lockutils [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.885278] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.885720] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Instance network_info: |[{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1083.886115] env[65107]: DEBUG oslo_concurrency.lockutils [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.886289] env[65107]: DEBUG nova.network.neutron [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1083.887592] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:11:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3e0aae3-33d1-403b-bfaf-306f77a1422e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47322e57-eaa6-4bb8-98ec-3098749bd52c', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.895839] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1083.900101] env[65107]: WARNING neutronclient.v2_0.client [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.900906] env[65107]: WARNING openstack [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.901472] env[65107]: WARNING openstack [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.910905] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1083.912134] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e33d526-58f5-4c40-9b67-dcfe59531606 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.934908] env[65107]: DEBUG nova.network.neutron [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1083.944146] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.944146] env[65107]: value = "task-5103481" [ 1083.944146] env[65107]: _type = "Task" [ 1083.944146] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.955614] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103481, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.958276] env[65107]: DEBUG nova.network.neutron [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Successfully updated port: 38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1084.041752] env[65107]: WARNING openstack [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.042241] env[65107]: WARNING openstack [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.066394] env[65107]: DEBUG oslo_vmware.api [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103480, 'name': PowerOffVM_Task, 'duration_secs': 0.31653} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.066869] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.067183] env[65107]: DEBUG nova.compute.manager [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1084.068068] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48506016-b7c4-4dfb-aa5c-40a1a5708a17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.084730] env[65107]: WARNING openstack [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.084992] env[65107]: WARNING openstack [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.099136] env[65107]: WARNING neutronclient.v2_0.client [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.099902] env[65107]: WARNING openstack [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.100884] env[65107]: WARNING openstack [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.151152] env[65107]: WARNING neutronclient.v2_0.client [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.151801] env[65107]: WARNING openstack [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.152150] env[65107]: WARNING openstack [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.210083] env[65107]: DEBUG nova.network.neutron [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Updating instance_info_cache with network_info: [{"id": "5343cf6f-59a3-4f9c-a534-449b9816168a", "address": "fa:16:3e:db:32:b1", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5343cf6f-59", "ovs_interfaceid": "5343cf6f-59a3-4f9c-a534-449b9816168a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1084.214109] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.215020] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.261511] env[65107]: DEBUG nova.network.neutron [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updated VIF entry in instance network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1084.262080] env[65107]: DEBUG nova.network.neutron [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1084.293574] env[65107]: DEBUG nova.network.neutron [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Port 9e9a520d-8984-4d4d-934b-f46dbdac5f4d binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1084.338730] env[65107]: DEBUG nova.compute.utils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1084.343442] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1084.344134] env[65107]: DEBUG nova.network.neutron [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1084.344134] env[65107]: WARNING neutronclient.v2_0.client [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.344417] env[65107]: WARNING neutronclient.v2_0.client [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.345017] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.345470] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.400870] env[65107]: DEBUG nova.policy [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1084.438912] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.456965] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103481, 'name': CreateVM_Task, 'duration_secs': 0.440722} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.459773] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1084.460788] env[65107]: WARNING neutronclient.v2_0.client [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.461325] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.461592] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.462087] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1084.462861] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.463052] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.463322] env[65107]: DEBUG nova.network.neutron [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1084.465248] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-369fbbff-e815-4474-baea-b5fc7b7cbbae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.474908] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1084.474908] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5227f509-8a29-4a52-6ce6-a6f115a628cb" [ 1084.474908] env[65107]: _type = "Task" [ 1084.474908] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.485024] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5227f509-8a29-4a52-6ce6-a6f115a628cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.581775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.560s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.616781] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222517d8-01fa-402c-8f3e-a1865b5a9108 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.625367] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4d923f-020d-4579-a8d4-f6f29f58b75a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.659098] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801b3447-d67e-4b55-96bc-3fae5e2f644a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.668407] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50291268-c200-4a77-9fcf-23bf616b1e6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.682765] env[65107]: DEBUG nova.compute.provider_tree [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.717465] env[65107]: DEBUG nova.network.neutron [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Successfully created port: 8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1084.720235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "refresh_cache-23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.766115] env[65107]: DEBUG oslo_concurrency.lockutils [req-301467ad-ec4b-4673-a8a8-48ea2b5d2ef9 req-cb0c0969-4469-4b52-89f2-c5c5c1a3b805 service nova] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.844678] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1084.945162] env[65107]: DEBUG nova.compute.manager [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65107) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1084.969400] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.969828] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.986691] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5227f509-8a29-4a52-6ce6-a6f115a628cb, 'name': SearchDatastore_Task, 'duration_secs': 0.012149} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.987034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.987891] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.987891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.987891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.987891] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.988086] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9345d7e-eba1-4c80-9800-17cf2489603c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.998586] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.998586] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1084.999312] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cc50a21-24f1-46a0-9295-6706cd48d2a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.006041] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1085.006041] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5225eaf3-fd4c-bea3-6574-038178716a7f" [ 1085.006041] env[65107]: _type = "Task" [ 1085.006041] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.018752] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5225eaf3-fd4c-bea3-6574-038178716a7f, 'name': SearchDatastore_Task, 'duration_secs': 0.010714} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.019720] env[65107]: DEBUG nova.network.neutron [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1085.022562] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54533f21-a667-44cc-a189-5e32cf887e6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.029033] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1085.029033] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7f34b-c2a8-def3-5627-ed8de1239f7c" [ 1085.029033] env[65107]: _type = "Task" [ 1085.029033] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.041993] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7f34b-c2a8-def3-5627-ed8de1239f7c, 'name': SearchDatastore_Task, 'duration_secs': 0.010414} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.042877] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.043260] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.049541] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.049793] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d/709eaf30-f1a6-49c2-a0fd-a47eee8aef2d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1085.050313] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4397fa91-71bf-463f-8706-abdff1049be2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.058129] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1085.058129] env[65107]: value = "task-5103482" [ 1085.058129] env[65107]: _type = "Task" [ 1085.058129] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.067235] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.090183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.158022] env[65107]: WARNING neutronclient.v2_0.client [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1085.158022] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.158988] env[65107]: WARNING openstack [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.186101] env[65107]: DEBUG nova.scheduler.client.report [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1085.207770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "8dbed811-3d4e-4cae-9981-0334801013aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.207770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.207770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "8dbed811-3d4e-4cae-9981-0334801013aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.207770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.207770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.209655] env[65107]: INFO nova.compute.manager [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Terminating instance [ 1085.318616] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.318864] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.319057] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.367991] env[65107]: DEBUG nova.network.neutron [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [{"id": "38203ad5-ce82-479b-8bb4-774b4fb04634", "address": "fa:16:3e:0c:be:68", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38203ad5-ce", "ovs_interfaceid": "38203ad5-ce82-479b-8bb4-774b4fb04634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.574605] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103482, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502807} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.576026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d/709eaf30-f1a6-49c2-a0fd-a47eee8aef2d.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.576421] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.576746] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22869ec6-0c5c-4673-9f65-03791ac87cb2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.588303] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1085.588303] env[65107]: value = "task-5103483" [ 1085.588303] env[65107]: _type = "Task" [ 1085.588303] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.601501] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.692335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.695164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.605s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.695369] env[65107]: DEBUG nova.objects.instance [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1085.714846] env[65107]: DEBUG nova.compute.manager [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1085.715052] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1085.716061] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd05a5fb-47a7-4e63-a7db-cf3edc185cd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.720495] env[65107]: INFO nova.scheduler.client.report [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Deleted allocations for instance f6536493-fb80-4fb1-9ee1-e530734e9adf [ 1085.728415] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.728714] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2cee27a-db7a-4527-9d87-ce414b346368 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.830437] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.830666] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.830851] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore2] 8dbed811-3d4e-4cae-9981-0334801013aa {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.831183] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae543045-31ee-4f87-a6ac-7d01dd779a2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.840498] env[65107]: DEBUG oslo_vmware.api [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1085.840498] env[65107]: value = "task-5103485" [ 1085.840498] env[65107]: _type = "Task" [ 1085.840498] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.851946] env[65107]: DEBUG oslo_vmware.api [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103485, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.855030] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1085.872292] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.872292] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Instance network_info: |[{"id": "38203ad5-ce82-479b-8bb4-774b4fb04634", "address": "fa:16:3e:0c:be:68", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38203ad5-ce", "ovs_interfaceid": "38203ad5-ce82-479b-8bb4-774b4fb04634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1085.873198] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:be:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38203ad5-ce82-479b-8bb4-774b4fb04634', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.881352] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1085.885035] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1085.886377] env[65107]: DEBUG nova.compute.manager [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Received event network-vif-plugged-38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1085.886613] env[65107]: DEBUG oslo_concurrency.lockutils [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Acquiring lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.886825] env[65107]: DEBUG oslo_concurrency.lockutils [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.886991] env[65107]: DEBUG oslo_concurrency.lockutils [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.887170] env[65107]: DEBUG nova.compute.manager [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] No waiting events found dispatching network-vif-plugged-38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1085.887336] env[65107]: WARNING nova.compute.manager [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Received unexpected event network-vif-plugged-38203ad5-ce82-479b-8bb4-774b4fb04634 for instance with vm_state building and task_state spawning. [ 1085.887491] env[65107]: DEBUG nova.compute.manager [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Received event network-changed-38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1085.887640] env[65107]: DEBUG nova.compute.manager [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Refreshing instance network info cache due to event network-changed-38203ad5-ce82-479b-8bb4-774b4fb04634. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1085.887821] env[65107]: DEBUG oslo_concurrency.lockutils [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Acquiring lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.887954] env[65107]: DEBUG oslo_concurrency.lockutils [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Acquired lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1085.888137] env[65107]: DEBUG nova.network.neutron [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Refreshing network info cache for port 38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1085.890872] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d433129-5327-42ad-bf22-753256ceaa95 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.908739] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1085.909161] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1085.909161] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1085.909339] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1085.909494] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1085.909638] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1085.909843] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.909999] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1085.910193] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1085.910379] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1085.910560] env[65107]: DEBUG nova.virt.hardware [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1085.911837] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325bf790-c1e4-4b8d-9dcc-c212e3081c16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.921872] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d41a0b0-6fdf-4d54-96c0-f78f56623b0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.925823] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1085.925823] env[65107]: value = "task-5103486" [ 1085.925823] env[65107]: _type = "Task" [ 1085.925823] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.944459] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103486, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.068411] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.098796] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077589} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.098940] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.099646] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c34cf9-fff6-4183-b82f-c4479575a522 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.122474] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d/709eaf30-f1a6-49c2-a0fd-a47eee8aef2d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.122802] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-499b427e-37f0-478a-b6b1-2870bad4f84c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.145239] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1086.145239] env[65107]: value = "task-5103487" [ 1086.145239] env[65107]: _type = "Task" [ 1086.145239] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.156280] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103487, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.160049] env[65107]: DEBUG nova.compute.manager [req-052e224f-a094-45cf-8780-f6d59ec6e49a req-261b56a9-019c-486b-bd79-4ea824df8f28 service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Received event network-vif-plugged-8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1086.160306] env[65107]: DEBUG oslo_concurrency.lockutils [req-052e224f-a094-45cf-8780-f6d59ec6e49a req-261b56a9-019c-486b-bd79-4ea824df8f28 service nova] Acquiring lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.160538] env[65107]: DEBUG oslo_concurrency.lockutils [req-052e224f-a094-45cf-8780-f6d59ec6e49a req-261b56a9-019c-486b-bd79-4ea824df8f28 service nova] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.160718] env[65107]: DEBUG oslo_concurrency.lockutils [req-052e224f-a094-45cf-8780-f6d59ec6e49a req-261b56a9-019c-486b-bd79-4ea824df8f28 service nova] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.160950] env[65107]: DEBUG nova.compute.manager [req-052e224f-a094-45cf-8780-f6d59ec6e49a req-261b56a9-019c-486b-bd79-4ea824df8f28 service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] No waiting events found dispatching network-vif-plugged-8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1086.161190] env[65107]: WARNING nova.compute.manager [req-052e224f-a094-45cf-8780-f6d59ec6e49a req-261b56a9-019c-486b-bd79-4ea824df8f28 service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Received unexpected event network-vif-plugged-8c74ade8-3d3b-4c53-ae8e-85d8e4177ced for instance with vm_state building and task_state spawning. [ 1086.228960] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b9c00e2-6161-4508-a088-30aa1f616513 tempest-ImagesNegativeTestJSON-1955032755 tempest-ImagesNegativeTestJSON-1955032755-project-member] Lock "f6536493-fb80-4fb1-9ee1-e530734e9adf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.256s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.262063] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.262063] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15d5d25d-cd9c-406b-b0ba-b07adfe2e87e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.272014] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1086.272014] env[65107]: value = "task-5103488" [ 1086.272014] env[65107]: _type = "Task" [ 1086.272014] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.279045] env[65107]: DEBUG nova.network.neutron [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Successfully updated port: 8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1086.284040] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103488, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.326525] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.352907] env[65107]: DEBUG oslo_vmware.api [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103485, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155465} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.353228] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.353528] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1086.353724] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1086.354172] env[65107]: INFO nova.compute.manager [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1086.354467] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1086.354705] env[65107]: DEBUG nova.compute.manager [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1086.354798] env[65107]: DEBUG nova.network.neutron [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1086.355082] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.356210] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.356210] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.364678] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.364871] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.365102] env[65107]: DEBUG nova.network.neutron [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1086.400453] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.407319] env[65107]: WARNING neutronclient.v2_0.client [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.407998] env[65107]: WARNING openstack [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.409086] env[65107]: WARNING openstack [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.437990] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103486, 'name': CreateVM_Task, 'duration_secs': 0.375462} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.437990] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1086.438221] env[65107]: WARNING neutronclient.v2_0.client [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.439280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.439280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.439280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1086.439776] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9837103-852e-46aa-acd9-cc568f0fde4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.447645] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1086.447645] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52899dfd-44f4-9210-8205-167d8ab953b9" [ 1086.447645] env[65107]: _type = "Task" [ 1086.447645] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.454857] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52899dfd-44f4-9210-8205-167d8ab953b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.543216] env[65107]: WARNING openstack [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.543621] env[65107]: WARNING openstack [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.642973] env[65107]: WARNING neutronclient.v2_0.client [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.644493] env[65107]: WARNING openstack [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.644887] env[65107]: WARNING openstack [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.654919] env[65107]: DEBUG nova.network.neutron [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1086.668461] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103487, 'name': ReconfigVM_Task, 'duration_secs': 0.341194} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.668461] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d/709eaf30-f1a6-49c2-a0fd-a47eee8aef2d.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.669338] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48fe4af3-ee99-4839-a930-5b64323873b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.677469] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1086.677469] env[65107]: value = "task-5103489" [ 1086.677469] env[65107]: _type = "Task" [ 1086.677469] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.687122] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103489, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.704748] env[65107]: DEBUG oslo_concurrency.lockutils [None req-aa691bc1-710a-4f15-8961-2494c35d414c tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.706106] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.638s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.744747] env[65107]: DEBUG nova.network.neutron [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updated VIF entry in instance network info cache for port 38203ad5-ce82-479b-8bb4-774b4fb04634. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1086.745156] env[65107]: DEBUG nova.network.neutron [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [{"id": "38203ad5-ce82-479b-8bb4-774b4fb04634", "address": "fa:16:3e:0c:be:68", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38203ad5-ce", "ovs_interfaceid": "38203ad5-ce82-479b-8bb4-774b4fb04634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1086.784580] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103488, 'name': PowerOffVM_Task, 'duration_secs': 0.391663} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.785183] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-b0b3e29e-9529-4d6d-b27c-c9d0169c5955" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.785345] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-b0b3e29e-9529-4d6d-b27c-c9d0169c5955" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.785518] env[65107]: DEBUG nova.network.neutron [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1086.786748] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1086.788134] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7e3d40-0d7f-4456-82a9-5b3f0bbe61b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.810817] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6c2a05-dc03-4c28-b2ca-348ead73ac8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.848390] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.848751] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6db4342b-52cd-4814-be7d-9bcf8e636fdb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.857079] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1086.857079] env[65107]: value = "task-5103490" [ 1086.857079] env[65107]: _type = "Task" [ 1086.857079] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.867988] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.868742] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.869054] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.876175] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103490, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.957024] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52899dfd-44f4-9210-8205-167d8ab953b9, 'name': SearchDatastore_Task, 'duration_secs': 0.017668} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.957024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1086.957274] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1086.957535] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.957682] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.957921] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1086.963472] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a55bfda-7676-413e-9def-f48b87d552ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.973632] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1086.973822] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1086.974596] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9928a56e-7e49-4326-83e3-68b392a25816 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.980620] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1086.980620] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f87980-057b-b047-6989-409c0b62cdcc" [ 1086.980620] env[65107]: _type = "Task" [ 1086.980620] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.985270] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.985654] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.998548] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f87980-057b-b047-6989-409c0b62cdcc, 'name': SearchDatastore_Task, 'duration_secs': 0.010629} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.999384] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e10d7d23-c9c0-4849-8459-dbafeb1acd24 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.005846] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1087.005846] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ce4cfb-3395-4b0f-563b-b20bb17d2582" [ 1087.005846] env[65107]: _type = "Task" [ 1087.005846] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.015281] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ce4cfb-3395-4b0f-563b-b20bb17d2582, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.052284] env[65107]: WARNING neutronclient.v2_0.client [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.052989] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.053354] env[65107]: WARNING openstack [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.158309] env[65107]: INFO nova.compute.manager [-] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Took 0.80 seconds to deallocate network for instance. [ 1087.187859] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103489, 'name': Rename_Task, 'duration_secs': 0.160061} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.188034] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.188765] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8c0d588-f9b4-41f2-9a63-c950d5945e46 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.195726] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1087.195726] env[65107]: value = "task-5103491" [ 1087.195726] env[65107]: _type = "Task" [ 1087.195726] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.198382] env[65107]: DEBUG nova.network.neutron [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [{"id": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "address": "fa:16:3e:44:a8:f3", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e9a520d-89", "ovs_interfaceid": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.210987] env[65107]: DEBUG nova.objects.instance [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'migration_context' on Instance uuid 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.212431] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103491, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.248252] env[65107]: DEBUG oslo_concurrency.lockutils [req-8baa987d-7d94-49c2-8049-18dc4150112a req-fb5f7f9c-e458-4ed2-9897-c10c8c92ec0e service nova] Releasing lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.293196] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.293646] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.339606] env[65107]: DEBUG nova.network.neutron [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1087.374474] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1087.375108] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.375621] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.375944] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.376325] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.376816] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-596c7d89-4c48-4ec2-bfd4-b3b21162cac6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.385443] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.386198] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.399556] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.399909] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.401100] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c67b156d-e5e5-4ffd-884c-13cc8e94cf3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.410491] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1087.410491] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52736b45-235a-d36a-bdb0-2c5271fe62f1" [ 1087.410491] env[65107]: _type = "Task" [ 1087.410491] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.424663] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52736b45-235a-d36a-bdb0-2c5271fe62f1, 'name': SearchDatastore_Task, 'duration_secs': 0.01092} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.425640] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f89db576-4f2e-44ff-a3c2-18b417ee9567 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.436421] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1087.436421] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b1c176-e0b9-30d9-e438-4e6943b13d42" [ 1087.436421] env[65107]: _type = "Task" [ 1087.436421] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.446457] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b1c176-e0b9-30d9-e438-4e6943b13d42, 'name': SearchDatastore_Task, 'duration_secs': 0.011468} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.446733] env[65107]: DEBUG oslo_concurrency.lockutils [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.446993] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. {{(pid=65107) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1087.448284] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-776c1283-6a67-4d0f-afcc-01a656526a21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.456392] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1087.456392] env[65107]: value = "task-5103492" [ 1087.456392] env[65107]: _type = "Task" [ 1087.456392] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.469224] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.477337] env[65107]: WARNING neutronclient.v2_0.client [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.478146] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.478490] env[65107]: WARNING openstack [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.517803] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ce4cfb-3395-4b0f-563b-b20bb17d2582, 'name': SearchDatastore_Task, 'duration_secs': 0.011564} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.518144] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.518460] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691/3f544b3f-703a-49d5-b3e8-16672766f691.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1087.518746] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-baf7c312-15a0-4567-91ae-9e0ca65b6bfe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.527510] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1087.527510] env[65107]: value = "task-5103493" [ 1087.527510] env[65107]: _type = "Task" [ 1087.527510] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.537877] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.642922] env[65107]: DEBUG nova.network.neutron [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Updating instance_info_cache with network_info: [{"id": "8c74ade8-3d3b-4c53-ae8e-85d8e4177ced", "address": "fa:16:3e:3a:8d:dc", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c74ade8-3d", "ovs_interfaceid": "8c74ade8-3d3b-4c53-ae8e-85d8e4177ced", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.666216] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.707486] env[65107]: DEBUG oslo_concurrency.lockutils [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.724884] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103491, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.973080] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103492, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.018179] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c2564b-b598-4e39-a37a-1854fd992d69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.028847] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cb00ef-b0e9-4246-ad68-0f4b34eb95a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.043910] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103493, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.070506] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ad7a6e-f9f8-42ca-8f9f-19a7d80b29f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.080339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b6d49c-1cf3-412c-8d11-da151722aa71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.096960] env[65107]: DEBUG nova.compute.provider_tree [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.145800] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-b0b3e29e-9529-4d6d-b27c-c9d0169c5955" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.146473] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Instance network_info: |[{"id": "8c74ade8-3d3b-4c53-ae8e-85d8e4177ced", "address": "fa:16:3e:3a:8d:dc", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c74ade8-3d", "ovs_interfaceid": "8c74ade8-3d3b-4c53-ae8e-85d8e4177ced", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1088.147153] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:8d:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c74ade8-3d3b-4c53-ae8e-85d8e4177ced', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.155158] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1088.155460] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.155774] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04db1725-f7bb-4190-aca0-e304d4f86ae0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.183431] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.183431] env[65107]: value = "task-5103494" [ 1088.183431] env[65107]: _type = "Task" [ 1088.183431] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.195738] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103494, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.221797] env[65107]: DEBUG oslo_vmware.api [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103491, 'name': PowerOnVM_Task, 'duration_secs': 0.556306} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.225361] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.225670] env[65107]: INFO nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Took 7.22 seconds to spawn the instance on the hypervisor. [ 1088.226168] env[65107]: DEBUG nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1088.229839] env[65107]: DEBUG nova.compute.manager [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Received event network-changed-8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1088.230310] env[65107]: DEBUG nova.compute.manager [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Refreshing instance network info cache due to event network-changed-8c74ade8-3d3b-4c53-ae8e-85d8e4177ced. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1088.230733] env[65107]: DEBUG oslo_concurrency.lockutils [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Acquiring lock "refresh_cache-b0b3e29e-9529-4d6d-b27c-c9d0169c5955" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.231084] env[65107]: DEBUG oslo_concurrency.lockutils [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Acquired lock "refresh_cache-b0b3e29e-9529-4d6d-b27c-c9d0169c5955" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1088.231420] env[65107]: DEBUG nova.network.neutron [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Refreshing network info cache for port 8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1088.238643] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb075f6-109a-4413-b997-6fae311bfa78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.269685] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87d846b-a2e6-40ec-99cd-3ebdbe7cb465 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.300958] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87908127-f4b0-4ee9-b387-64e7a5c88a44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.311395] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1088.468196] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103492, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591847} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.468347] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. [ 1088.469155] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf9eb94-4be7-489a-b45e-35da67ae42b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.496885] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.496885] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec9e430a-27c3-4ef0-9fb3-a6fab77936b6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.517728] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1088.517728] env[65107]: value = "task-5103495" [ 1088.517728] env[65107]: _type = "Task" [ 1088.517728] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.527915] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103495, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.543472] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.776659} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.543818] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691/3f544b3f-703a-49d5-b3e8-16672766f691.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1088.544060] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1088.544344] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10bdc3f2-bf45-4252-a486-dcbf92f9a0e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.555420] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1088.555420] env[65107]: value = "task-5103496" [ 1088.555420] env[65107]: _type = "Task" [ 1088.555420] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.566032] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.600328] env[65107]: DEBUG nova.scheduler.client.report [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1088.694996] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103494, 'name': CreateVM_Task, 'duration_secs': 0.409358} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.695270] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1088.695882] env[65107]: WARNING neutronclient.v2_0.client [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.696327] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.696495] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1088.696840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1088.697132] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efcfa98f-8c1f-4469-af8e-50663711d97c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.703238] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1088.703238] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527e4d9b-e002-aad2-35f5-db949a7ec7bf" [ 1088.703238] env[65107]: _type = "Task" [ 1088.703238] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.713088] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527e4d9b-e002-aad2-35f5-db949a7ec7bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.743059] env[65107]: WARNING neutronclient.v2_0.client [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.743441] env[65107]: WARNING openstack [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.743818] env[65107]: WARNING openstack [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.765601] env[65107]: INFO nova.compute.manager [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Took 17.77 seconds to build instance. [ 1088.818627] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.818978] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19da8f91-679b-4e44-aad6-7c8b641e4902 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.829473] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1088.829473] env[65107]: value = "task-5103497" [ 1088.829473] env[65107]: _type = "Task" [ 1088.829473] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.855358] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.905026] env[65107]: WARNING openstack [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.905412] env[65107]: WARNING openstack [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1089.022710] env[65107]: WARNING neutronclient.v2_0.client [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.024260] env[65107]: WARNING openstack [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1089.024260] env[65107]: WARNING openstack [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1089.036330] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.036536] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.041465] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103495, 'name': ReconfigVM_Task, 'duration_secs': 0.350882} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.042026] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.042889] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1d6eae-c69c-4829-be0f-43c56b55048b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.074292] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abe53a3c-cd63-47b5-9d52-bfd5f24b4091 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.091022] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075331} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.092491] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.092863] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1089.092863] env[65107]: value = "task-5103498" [ 1089.092863] env[65107]: _type = "Task" [ 1089.092863] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.093813] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e44771-c92e-4cb8-bc42-59478741104b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.127360] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691/3f544b3f-703a-49d5-b3e8-16672766f691.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.138179] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af7d263c-614d-43f1-b0b6-ab7d8f1cbf9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.154279] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103498, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.162482] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1089.162482] env[65107]: value = "task-5103499" [ 1089.162482] env[65107]: _type = "Task" [ 1089.162482] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.174484] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103499, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.215626] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527e4d9b-e002-aad2-35f5-db949a7ec7bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010871} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.216238] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1089.216303] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.216647] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.216837] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1089.217027] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.217343] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec7b6138-239e-4aa3-99b9-ce3b44f2af4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.228398] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.228631] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1089.229556] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08ef583e-84a6-4f86-b3d3-0100971012a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.238128] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1089.238128] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52786042-74c3-8900-e0df-3adce184af41" [ 1089.238128] env[65107]: _type = "Task" [ 1089.238128] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.248315] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52786042-74c3-8900-e0df-3adce184af41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.267635] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c16b869d-6e87-4e39-b86d-7dc52647ee44 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.288s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.297658] env[65107]: DEBUG nova.network.neutron [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Updated VIF entry in instance network info cache for port 8c74ade8-3d3b-4c53-ae8e-85d8e4177ced. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1089.298146] env[65107]: DEBUG nova.network.neutron [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Updating instance_info_cache with network_info: [{"id": "8c74ade8-3d3b-4c53-ae8e-85d8e4177ced", "address": "fa:16:3e:3a:8d:dc", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c74ade8-3d", "ovs_interfaceid": "8c74ade8-3d3b-4c53-ae8e-85d8e4177ced", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1089.340492] env[65107]: DEBUG oslo_vmware.api [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103497, 'name': PowerOnVM_Task, 'duration_secs': 0.478804} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.340757] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.340941] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-394e97c6-1ade-4fbd-9e6c-709ffda8b010 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance '7373fbbe-52c1-4955-8da9-a5563d71391b' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.542769] env[65107]: INFO nova.compute.manager [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Detaching volume e8465018-2a5c-45af-b181-4202b8bd8fce [ 1089.604915] env[65107]: INFO nova.virt.block_device [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Attempting to driver detach volume e8465018-2a5c-45af-b181-4202b8bd8fce from mountpoint /dev/sdb [ 1089.605127] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.605325] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992831', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'name': 'volume-e8465018-2a5c-45af-b181-4202b8bd8fce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e12cb11f-7d6f-47d6-a31e-bf2654174d38', 'attached_at': '', 'detached_at': '', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'serial': 'e8465018-2a5c-45af-b181-4202b8bd8fce'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.606300] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec83dd48-d992-4339-8880-1780340e353d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.616126] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103498, 'name': ReconfigVM_Task, 'duration_secs': 0.445829} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.633042] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.633435] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31eab613-49a2-4927-971c-bd2f753d6cbe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.636074] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67075062-1157-47de-96e7-fde25871891d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.640374] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.934s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.646208] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.980s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.646436] env[65107]: DEBUG nova.objects.instance [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'resources' on Instance uuid 8dbed811-3d4e-4cae-9981-0334801013aa {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.657617] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1089.657617] env[65107]: value = "task-5103500" [ 1089.657617] env[65107]: _type = "Task" [ 1089.657617] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.659697] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0441d2e3-347c-4648-abb6-cf819d451085 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.697256] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8c5c03-f8aa-4236-92c3-2496ce4de653 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.702367] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103500, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.704894] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.720711] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] The volume has not been displaced from its original location: [datastore1] volume-e8465018-2a5c-45af-b181-4202b8bd8fce/volume-e8465018-2a5c-45af-b181-4202b8bd8fce.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.726631] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.727178] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9add80f-306b-4cf6-b069-93d8c4511783 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.751114] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52786042-74c3-8900-e0df-3adce184af41, 'name': SearchDatastore_Task, 'duration_secs': 0.017367} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.752074] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-021e1538-0078-4425-8cbd-eefa90310ca4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.756779] env[65107]: DEBUG oslo_vmware.api [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1089.756779] env[65107]: value = "task-5103501" [ 1089.756779] env[65107]: _type = "Task" [ 1089.756779] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.762479] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1089.762479] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dc4bcf-8ee3-490b-01e7-15ad33109039" [ 1089.762479] env[65107]: _type = "Task" [ 1089.762479] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.772036] env[65107]: DEBUG oslo_vmware.api [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103501, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.776465] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dc4bcf-8ee3-490b-01e7-15ad33109039, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.801743] env[65107]: DEBUG oslo_concurrency.lockutils [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] Releasing lock "refresh_cache-b0b3e29e-9529-4d6d-b27c-c9d0169c5955" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1089.802044] env[65107]: DEBUG nova.compute.manager [req-364093b6-a5e0-44c6-932d-2cd96c960799 req-14a981d0-539c-4de0-b4c8-1f6c885316fd service nova] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Received event network-vif-deleted-6091d1cb-29b9-4a22-8827-990e8ea7d318 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1090.181805] env[65107]: DEBUG oslo_vmware.api [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103500, 'name': PowerOnVM_Task, 'duration_secs': 0.473188} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.182870] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103499, 'name': ReconfigVM_Task, 'duration_secs': 0.774315} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.186260] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.189102] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691/3f544b3f-703a-49d5-b3e8-16672766f691.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.192743] env[65107]: DEBUG nova.compute.manager [None req-af249b6e-a784-4918-bc83-f5379dbfab9b tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1090.193265] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65c56273-77fe-40a6-b877-d37a6cf4d036 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.198024] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5dd371-0f0f-49bb-ab7a-592d22cd2c29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.215571] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1090.215571] env[65107]: value = "task-5103502" [ 1090.215571] env[65107]: _type = "Task" [ 1090.215571] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.231076] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103502, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.279426] env[65107]: DEBUG oslo_vmware.api [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103501, 'name': ReconfigVM_Task, 'duration_secs': 0.299219} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.283841] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.296493] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8ddbf1e-2bb4-44d4-ab14-252d51654de7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.310618] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dc4bcf-8ee3-490b-01e7-15ad33109039, 'name': SearchDatastore_Task, 'duration_secs': 0.01383} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.313555] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1090.313856] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] b0b3e29e-9529-4d6d-b27c-c9d0169c5955/b0b3e29e-9529-4d6d-b27c-c9d0169c5955.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.315297] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7e1bce1-2fda-4134-9f5e-507c0e025e7c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.321959] env[65107]: DEBUG oslo_vmware.api [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1090.321959] env[65107]: value = "task-5103503" [ 1090.321959] env[65107]: _type = "Task" [ 1090.321959] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.324575] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1090.324575] env[65107]: value = "task-5103504" [ 1090.324575] env[65107]: _type = "Task" [ 1090.324575] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.340399] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.344357] env[65107]: DEBUG oslo_vmware.api [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.469357] env[65107]: DEBUG nova.compute.manager [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1090.469357] env[65107]: DEBUG nova.compute.manager [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing instance network info cache due to event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1090.469357] env[65107]: DEBUG oslo_concurrency.lockutils [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.469357] env[65107]: DEBUG oslo_concurrency.lockutils [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.469357] env[65107]: DEBUG nova.network.neutron [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1090.523169] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acca43b0-da98-4e1d-a921-0ae225c9d33b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.535903] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d260b34-d260-4261-92eb-0ecb8a64d678 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.572856] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd03003-5cc5-45e6-a3aa-e645ccd7503c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.584852] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d858332-b3ba-4c0a-a8a3-c3119bb302fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.603035] env[65107]: DEBUG nova.compute.provider_tree [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.738591] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103502, 'name': Rename_Task, 'duration_secs': 0.177357} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.738975] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1090.739155] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-330aed10-004e-4fc9-be04-c9575a577247 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.751791] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1090.751791] env[65107]: value = "task-5103505" [ 1090.751791] env[65107]: _type = "Task" [ 1090.751791] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.769696] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.841313] env[65107]: DEBUG oslo_vmware.api [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103503, 'name': ReconfigVM_Task, 'duration_secs': 0.1856} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.844988] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992831', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'name': 'volume-e8465018-2a5c-45af-b181-4202b8bd8fce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e12cb11f-7d6f-47d6-a31e-bf2654174d38', 'attached_at': '', 'detached_at': '', 'volume_id': 'e8465018-2a5c-45af-b181-4202b8bd8fce', 'serial': 'e8465018-2a5c-45af-b181-4202b8bd8fce'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1090.847673] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103504, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.971615] env[65107]: WARNING neutronclient.v2_0.client [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1090.972412] env[65107]: WARNING openstack [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1090.972773] env[65107]: WARNING openstack [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.080529] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.080770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.106204] env[65107]: DEBUG nova.scheduler.client.report [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1091.195766] env[65107]: INFO nova.compute.manager [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Swapping old allocation on dict_keys(['8adccf8d-6d73-43c0-ad15-ee89967ab8cd']) held by migration f9d5050a-6b81-4c45-ae1d-c9b5fa915cc2 for instance [ 1091.226527] env[65107]: DEBUG nova.scheduler.client.report [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Overwriting current allocation {'allocations': {'8adccf8d-6d73-43c0-ad15-ee89967ab8cd': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 146}}, 'project_id': 'f379144b78764fe394039d87b043a946', 'user_id': '6590cb74e83e45bd944f23fa888a4b04', 'consumer_generation': 1} on consumer 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 {{(pid=65107) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1091.241748] env[65107]: WARNING openstack [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.243734] env[65107]: WARNING openstack [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.263541] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.300419] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.344021] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.642069} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.344021] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] b0b3e29e-9529-4d6d-b27c-c9d0169c5955/b0b3e29e-9529-4d6d-b27c-c9d0169c5955.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.344021] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.344021] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a12fc6c0-1b53-4c76-9336-85fbd92e59a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.353086] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1091.353086] env[65107]: value = "task-5103506" [ 1091.353086] env[65107]: _type = "Task" [ 1091.353086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.369292] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103506, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.401806] env[65107]: DEBUG nova.objects.instance [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'flavor' on Instance uuid e12cb11f-7d6f-47d6-a31e-bf2654174d38 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.437714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.437989] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.438228] env[65107]: DEBUG nova.network.neutron [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1091.526562] env[65107]: WARNING neutronclient.v2_0.client [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.527361] env[65107]: WARNING openstack [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.527709] env[65107]: WARNING openstack [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.584173] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1091.611407] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.650191] env[65107]: INFO nova.scheduler.client.report [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted allocations for instance 8dbed811-3d4e-4cae-9981-0334801013aa [ 1091.662738] env[65107]: DEBUG nova.network.neutron [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updated VIF entry in instance network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1091.666017] env[65107]: DEBUG nova.network.neutron [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1091.763556] env[65107]: DEBUG oslo_vmware.api [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103505, 'name': PowerOnVM_Task, 'duration_secs': 0.777241} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.764328] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1091.764328] env[65107]: INFO nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Took 8.33 seconds to spawn the instance on the hypervisor. [ 1091.764449] env[65107]: DEBUG nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1091.765300] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20daab83-2d83-425e-aea1-8bc91786ba66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.866176] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103506, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104864} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.867084] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1091.867347] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3397afb9-21d0-434e-8aee-1a7f7e8a1a20 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.893053] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b0b3e29e-9529-4d6d-b27c-c9d0169c5955/b0b3e29e-9529-4d6d-b27c-c9d0169c5955.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.894446] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb684323-7ab7-482c-843b-8e8939c5dea3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.922821] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1091.922821] env[65107]: value = "task-5103507" [ 1091.922821] env[65107]: _type = "Task" [ 1091.922821] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.935733] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103507, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.941181] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.942402] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.942973] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.089774] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.090544] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.117054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.117175] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.118889] env[65107]: INFO nova.compute.claims [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.160035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-33367b8d-504e-47e3-8643-27111f1d8a12 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "8dbed811-3d4e-4cae-9981-0334801013aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.953s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.164658] env[65107]: WARNING neutronclient.v2_0.client [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.165344] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.165746] env[65107]: WARNING openstack [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.173769] env[65107]: DEBUG oslo_concurrency.lockutils [req-79e01b86-9826-431c-9884-599ac3795212 req-6b77e733-1374-407c-893d-565cb43e3947 service nova] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.292379] env[65107]: DEBUG nova.network.neutron [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [{"id": "416a2590-6a42-4a32-a922-a9d5f95786d6", "address": "fa:16:3e:a7:c9:15", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416a2590-6a", "ovs_interfaceid": "416a2590-6a42-4a32-a922-a9d5f95786d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.294085] env[65107]: INFO nova.compute.manager [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Took 18.39 seconds to build instance. [ 1092.418868] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c066ebe6-bc7d-48d0-8db1-afb29ccc65bc tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.382s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.436417] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103507, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.562251] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.562251] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.693220] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.693884] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.792812] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.793075] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.796964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-8cddc4d1-c94c-4f2b-9c4b-3125e4638050" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.796964] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fc3160-7690-47ab-ba05-3653d7b27e06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.800172] env[65107]: DEBUG oslo_concurrency.lockutils [None req-914b69b6-393e-413b-98a2-795981939c54 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.909s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.805997] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073399d3-b49c-4fd4-a603-e06c84609853 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.935671] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103507, 'name': ReconfigVM_Task, 'duration_secs': 0.610214} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.936961] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b0b3e29e-9529-4d6d-b27c-c9d0169c5955/b0b3e29e-9529-4d6d-b27c-c9d0169c5955.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.936961] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb76c190-6409-4c70-90c6-cf1a2ec149d5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.939690] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "7373fbbe-52c1-4955-8da9-a5563d71391b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.940254] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.940556] env[65107]: DEBUG nova.compute.manager [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Going to confirm migration 6 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1092.947798] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1092.947798] env[65107]: value = "task-5103508" [ 1092.947798] env[65107]: _type = "Task" [ 1092.947798] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.960025] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103508, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.065684] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1093.197209] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1093.302532] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.302722] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.302869] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.303148] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.303632] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.303632] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.303710] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1093.303834] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.342578] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.342924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.450417] env[65107]: WARNING neutronclient.v2_0.client [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.463735] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103508, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.480851] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f0df8e-fb44-4f9e-be09-1de47cfffa3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.490778] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67c1e7d-1fd7-4cfa-b1af-3e2843026746 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.528019] env[65107]: WARNING neutronclient.v2_0.client [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.528481] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.528650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquired lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.528827] env[65107]: DEBUG nova.network.neutron [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1093.529022] env[65107]: DEBUG nova.objects.instance [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'info_cache' on Instance uuid 7373fbbe-52c1-4955-8da9-a5563d71391b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.530855] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a808371-0665-4c64-8400-72fb127b9f53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.540462] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f0f57b-7b80-4f5c-ba62-2b71323f08c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.556541] env[65107]: DEBUG nova.compute.provider_tree [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.586191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.723415] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.809941] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.846697] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1093.930285] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1093.930617] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-881dd895-b948-4a40-b912-48ae1298b69d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.939416] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1093.939416] env[65107]: value = "task-5103509" [ 1093.939416] env[65107]: _type = "Task" [ 1093.939416] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.948375] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103509, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.965376] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103508, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.059717] env[65107]: DEBUG nova.scheduler.client.report [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1094.373152] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.451401] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103509, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.463440] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103508, 'name': Rename_Task, 'duration_secs': 1.295299} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.463833] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1094.464191] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31f10b8f-b8e7-42bc-8f78-724f64c619a1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.470767] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1094.470767] env[65107]: value = "task-5103510" [ 1094.470767] env[65107]: _type = "Task" [ 1094.470767] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.479430] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.499289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.499735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.499919] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.500141] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.500370] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.502680] env[65107]: INFO nova.compute.manager [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Terminating instance [ 1094.538586] env[65107]: WARNING neutronclient.v2_0.client [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.539358] env[65107]: WARNING openstack [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.539731] env[65107]: WARNING openstack [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.565076] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.565603] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1094.569168] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.983s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.570632] env[65107]: INFO nova.compute.claims [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1094.659559] env[65107]: WARNING openstack [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.659928] env[65107]: WARNING openstack [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.731771] env[65107]: WARNING neutronclient.v2_0.client [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.732507] env[65107]: WARNING openstack [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.732903] env[65107]: WARNING openstack [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.816084] env[65107]: DEBUG nova.compute.manager [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1094.826747] env[65107]: DEBUG nova.network.neutron [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [{"id": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "address": "fa:16:3e:44:a8:f3", "network": {"id": "9a6a037a-f1d1-42e3-bc8d-85db4f6d9d49", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-419854694-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4778fe9152224fd29d1f6220a19b5a36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e9a520d-89", "ovs_interfaceid": "9e9a520d-8984-4d4d-934b-f46dbdac5f4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.950978] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103509, 'name': PowerOffVM_Task, 'duration_secs': 0.5166} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.951282] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1094.951940] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1094.952162] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1094.952316] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1094.952491] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1094.952803] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1094.952958] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1094.953192] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.953364] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1094.953532] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1094.953718] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1094.953893] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1094.959014] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaa57ca0-75b7-40e7-81aa-e05a746bdfcd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.978259] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1094.978259] env[65107]: value = "task-5103511" [ 1094.978259] env[65107]: _type = "Task" [ 1094.978259] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.985040] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103510, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.990446] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.007916] env[65107]: DEBUG nova.compute.manager [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1095.007916] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.007916] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d03856-96e1-4134-bc2e-1addb7d683ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.015858] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.016175] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bdc47f1-d498-4b38-ba5a-ce8271eece19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.023905] env[65107]: DEBUG oslo_vmware.api [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1095.023905] env[65107]: value = "task-5103512" [ 1095.023905] env[65107]: _type = "Task" [ 1095.023905] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.032914] env[65107]: DEBUG oslo_vmware.api [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103512, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.079716] env[65107]: DEBUG nova.compute.utils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1095.081846] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1095.082097] env[65107]: DEBUG nova.network.neutron [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1095.082465] env[65107]: WARNING neutronclient.v2_0.client [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.082838] env[65107]: WARNING neutronclient.v2_0.client [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.083543] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.083924] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.135156] env[65107]: DEBUG nova.policy [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1095.311607] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.311978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.329203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Releasing lock "refresh_cache-7373fbbe-52c1-4955-8da9-a5563d71391b" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.329203] env[65107]: DEBUG nova.objects.instance [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'migration_context' on Instance uuid 7373fbbe-52c1-4955-8da9-a5563d71391b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.341234] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.482558] env[65107]: DEBUG oslo_vmware.api [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103510, 'name': PowerOnVM_Task, 'duration_secs': 0.889101} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.486096] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.486096] env[65107]: INFO nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Took 9.63 seconds to spawn the instance on the hypervisor. [ 1095.486348] env[65107]: DEBUG nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1095.487191] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21766da1-e256-47b6-8e04-31db80e7b3d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.495872] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103511, 'name': ReconfigVM_Task, 'duration_secs': 0.230118} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.501256] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d078a7-90eb-4642-acf8-c339afccb9e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.525142] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1095.525405] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1095.525625] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1095.525780] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1095.525927] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1095.526085] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1095.526294] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.526446] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1095.526611] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1095.526777] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1095.526951] env[65107]: DEBUG nova.virt.hardware [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1095.530772] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af85648c-86be-499e-8d1d-757feacd0180 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.541049] env[65107]: DEBUG oslo_vmware.api [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103512, 'name': PowerOffVM_Task, 'duration_secs': 0.290007} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.541049] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1095.541049] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5231d100-5a8a-cf22-b98d-68793168cb44" [ 1095.541049] env[65107]: _type = "Task" [ 1095.541049] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.541049] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.541424] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.541687] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49cbe074-b083-4540-a125-4d73127e6204 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.548594] env[65107]: DEBUG nova.network.neutron [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Successfully created port: d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1095.560289] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5231d100-5a8a-cf22-b98d-68793168cb44, 'name': SearchDatastore_Task, 'duration_secs': 0.0202} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.566139] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1095.566463] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9213fc8e-6b95-4a23-b072-727211928db8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.586333] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1095.596621] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1095.596621] env[65107]: value = "task-5103514" [ 1095.596621] env[65107]: _type = "Task" [ 1095.596621] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.607522] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103514, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.633121] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1095.633308] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1095.633500] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleting the datastore file [datastore2] e12cb11f-7d6f-47d6-a31e-bf2654174d38 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.633884] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c05530c0-7940-46b2-86a7-9d93cca3b19a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.642336] env[65107]: DEBUG oslo_vmware.api [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1095.642336] env[65107]: value = "task-5103515" [ 1095.642336] env[65107]: _type = "Task" [ 1095.642336] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.655772] env[65107]: DEBUG oslo_vmware.api [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.815579] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1095.831912] env[65107]: DEBUG nova.objects.base [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Object Instance<7373fbbe-52c1-4955-8da9-a5563d71391b> lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1095.833209] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0feaf8-1bec-40b7-88c2-0f60380bb708 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.863210] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bec31f8c-86d7-40c0-be7f-48b42185c1cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.872229] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1095.872229] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c3fbd-9368-e1dc-595a-b022bacbaebf" [ 1095.872229] env[65107]: _type = "Task" [ 1095.872229] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.885853] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c3fbd-9368-e1dc-595a-b022bacbaebf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.933997] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f919689-ce17-4ae1-9fae-9d71b543cc53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.943188] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909a6cee-4d34-4ee4-b180-34ba5e0616e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.976037] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f738d4c6-b36f-4d59-83a1-b0201765990a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.984246] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4298b3fe-5556-413d-9bee-b65e0f6de65e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.999308] env[65107]: DEBUG nova.compute.provider_tree [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.013661] env[65107]: INFO nova.compute.manager [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Took 19.32 seconds to build instance. [ 1096.106582] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103514, 'name': ReconfigVM_Task, 'duration_secs': 0.284818} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.106946] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.107791] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc05e56-997e-4abd-996d-6cedc34c9baf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.136556] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.137821] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9238a203-1e04-4a8f-a9af-fd010306e692 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.160928] env[65107]: DEBUG oslo_vmware.api [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257442} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.162496] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.162716] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.162905] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.163121] env[65107]: INFO nova.compute.manager [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1096.163442] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1096.163798] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1096.163798] env[65107]: value = "task-5103516" [ 1096.163798] env[65107]: _type = "Task" [ 1096.163798] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.164043] env[65107]: DEBUG nova.compute.manager [-] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1096.164135] env[65107]: DEBUG nova.network.neutron [-] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1096.164404] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.164986] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.165283] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.183916] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.255397] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.342220] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.384155] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c3fbd-9368-e1dc-595a-b022bacbaebf, 'name': SearchDatastore_Task, 'duration_secs': 0.020534} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.384496] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.503575] env[65107]: DEBUG nova.scheduler.client.report [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.516052] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5a5a84b6-0373-4c60-b4cd-285fa8417934 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.828s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.598179] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1096.629246] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1096.629621] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1096.629816] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1096.630060] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1096.630357] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1096.630561] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1096.630807] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.631091] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1096.631224] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1096.631551] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1096.631778] env[65107]: DEBUG nova.virt.hardware [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1096.632688] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c88fc5f-488e-4d30-ba65-6ab5a7b3393c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.642750] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9623afd9-8393-496c-b789-27617b25d204 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.682811] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.009310] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.009862] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1097.012714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.289s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.015868] env[65107]: INFO nova.compute.claims [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.108858] env[65107]: DEBUG nova.network.neutron [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Successfully updated port: d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1097.165418] env[65107]: DEBUG nova.compute.manager [req-b431f5b3-2b19-4736-abdc-17521b25034b req-c094ad6e-dca7-4d8b-82a3-b2c5759fc9db service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Received event network-vif-deleted-ed79bc21-5c5f-44e5-ba58-55e43c814088 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1097.165997] env[65107]: INFO nova.compute.manager [req-b431f5b3-2b19-4736-abdc-17521b25034b req-c094ad6e-dca7-4d8b-82a3-b2c5759fc9db service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Neutron deleted interface ed79bc21-5c5f-44e5-ba58-55e43c814088; detaching it from the instance and deleting it from the info cache [ 1097.166072] env[65107]: DEBUG nova.network.neutron [req-b431f5b3-2b19-4736-abdc-17521b25034b req-c094ad6e-dca7-4d8b-82a3-b2c5759fc9db service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.186681] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103516, 'name': ReconfigVM_Task, 'duration_secs': 0.677658} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.187065] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050/8cddc4d1-c94c-4f2b-9c4b-3125e4638050.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.188756] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc060d8d-1623-4b8e-8ca2-0e0ccde98392 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.214398] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b499458-acc0-4933-b9c6-6c4f33e4a718 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.236562] env[65107]: DEBUG nova.network.neutron [-] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.239925] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5f4a20-a9df-4a9c-8622-d5a14b19895f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.265019] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d37da1-1ca8-4d00-9c00-c295f97cf534 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.273976] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1097.274291] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62ca6ca1-6733-43af-b459-e224c0116713 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.281924] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1097.281924] env[65107]: value = "task-5103517" [ 1097.281924] env[65107]: _type = "Task" [ 1097.281924] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.290765] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.341023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.341299] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.341581] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.341836] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.342065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.344943] env[65107]: INFO nova.compute.manager [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Terminating instance [ 1097.353073] env[65107]: DEBUG nova.compute.manager [req-d618c48a-fa70-4694-a90b-b154ec28a5a5 req-982a51df-9b4d-46eb-a9e3-ea93819ab2df service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-vif-plugged-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1097.353383] env[65107]: DEBUG oslo_concurrency.lockutils [req-d618c48a-fa70-4694-a90b-b154ec28a5a5 req-982a51df-9b4d-46eb-a9e3-ea93819ab2df service nova] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.353715] env[65107]: DEBUG oslo_concurrency.lockutils [req-d618c48a-fa70-4694-a90b-b154ec28a5a5 req-982a51df-9b4d-46eb-a9e3-ea93819ab2df service nova] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.353770] env[65107]: DEBUG oslo_concurrency.lockutils [req-d618c48a-fa70-4694-a90b-b154ec28a5a5 req-982a51df-9b4d-46eb-a9e3-ea93819ab2df service nova] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.353919] env[65107]: DEBUG nova.compute.manager [req-d618c48a-fa70-4694-a90b-b154ec28a5a5 req-982a51df-9b4d-46eb-a9e3-ea93819ab2df service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] No waiting events found dispatching network-vif-plugged-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1097.354085] env[65107]: WARNING nova.compute.manager [req-d618c48a-fa70-4694-a90b-b154ec28a5a5 req-982a51df-9b4d-46eb-a9e3-ea93819ab2df service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received unexpected event network-vif-plugged-d04b7a1c-d3b0-4211-b44b-2e39791f7497 for instance with vm_state building and task_state spawning. [ 1097.519495] env[65107]: DEBUG nova.compute.utils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1097.522882] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1097.523095] env[65107]: DEBUG nova.network.neutron [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1097.523466] env[65107]: WARNING neutronclient.v2_0.client [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.523820] env[65107]: WARNING neutronclient.v2_0.client [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.524437] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.524792] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.581418] env[65107]: DEBUG nova.policy [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6723c29f05774e138ee11fa945c058e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '139a38d3b7bd4d3a8861d2e1e6c7a2c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1097.613560] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.613764] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.613949] env[65107]: DEBUG nova.network.neutron [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1097.669155] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2a83f69-d760-406e-b397-d52a8c5763a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.681178] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5b696b-3375-4c47-9ea3-8d679d42d3c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.722943] env[65107]: DEBUG nova.compute.manager [req-b431f5b3-2b19-4736-abdc-17521b25034b req-c094ad6e-dca7-4d8b-82a3-b2c5759fc9db service nova] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Detach interface failed, port_id=ed79bc21-5c5f-44e5-ba58-55e43c814088, reason: Instance e12cb11f-7d6f-47d6-a31e-bf2654174d38 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1097.739089] env[65107]: INFO nova.compute.manager [-] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Took 1.57 seconds to deallocate network for instance. [ 1097.793845] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103517, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.849846] env[65107]: DEBUG nova.compute.manager [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1097.850393] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.850862] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad9d196-710a-4e2c-a378-b8816d1781cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.861217] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.861510] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0904e654-7e82-463d-9e8a-20406b7e9d44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.869536] env[65107]: DEBUG oslo_vmware.api [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1097.869536] env[65107]: value = "task-5103518" [ 1097.869536] env[65107]: _type = "Task" [ 1097.869536] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.882060] env[65107]: DEBUG oslo_vmware.api [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.922409] env[65107]: DEBUG nova.network.neutron [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Successfully created port: 429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1098.025165] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1098.117316] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.117735] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.163771] env[65107]: DEBUG nova.network.neutron [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1098.209373] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.209833] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.245772] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.297728] env[65107]: DEBUG oslo_vmware.api [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103517, 'name': PowerOnVM_Task, 'duration_secs': 0.678196} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.299227] env[65107]: WARNING neutronclient.v2_0.client [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.300125] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.300643] env[65107]: WARNING openstack [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.310041] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1098.382101] env[65107]: DEBUG oslo_vmware.api [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103518, 'name': PowerOffVM_Task, 'duration_secs': 0.302357} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.385129] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.385329] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.385831] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1509c855-3a0f-4b11-9203-bf2c9bbcc571 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.397170] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0864978f-961a-4f49-b6bf-14d7b3e27ec1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.405754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782a9ba5-ed88-4476-88d1-576f6d1d4ea5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.437249] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0637246-a481-4e26-9b17-a302a16a6f74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.445602] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174953b4-bc15-4ff5-8514-7fd432168c1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.454748] env[65107]: DEBUG nova.network.neutron [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.465342] env[65107]: DEBUG nova.compute.provider_tree [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.468859] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.469074] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.469253] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore1] b0b3e29e-9529-4d6d-b27c-c9d0169c5955 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.470102] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f600361-5c58-4eaa-a029-07bdf7837569 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.478028] env[65107]: DEBUG oslo_vmware.api [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1098.478028] env[65107]: value = "task-5103520" [ 1098.478028] env[65107]: _type = "Task" [ 1098.478028] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.487412] env[65107]: DEBUG oslo_vmware.api [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.851463] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.851769] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.970527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.973036] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Instance network_info: |[{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1098.973036] env[65107]: DEBUG nova.scheduler.client.report [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.975224] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:eb:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3e0aae3-33d1-403b-bfaf-306f77a1422e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd04b7a1c-d3b0-4211-b44b-2e39791f7497', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1098.983306] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1098.984137] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1098.988050] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-473bddfd-0c9a-4139-86bf-068eee455ddb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.008928] env[65107]: DEBUG oslo_vmware.api [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171467} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.010380] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.010588] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1099.010874] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1099.010943] env[65107]: INFO nova.compute.manager [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1099.011192] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1099.011387] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1099.011387] env[65107]: value = "task-5103521" [ 1099.011387] env[65107]: _type = "Task" [ 1099.011387] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.011588] env[65107]: DEBUG nova.compute.manager [-] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1099.011683] env[65107]: DEBUG nova.network.neutron [-] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1099.011928] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.012464] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.012720] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.030276] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103521, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.034877] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1099.052541] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.064407] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1099.064407] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1099.064407] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1099.064633] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1099.064749] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1099.064919] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1099.065125] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.065390] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1099.065598] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1099.065938] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1099.065938] env[65107]: DEBUG nova.virt.hardware [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1099.066838] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c22efb-5641-42c4-b6e1-437c22722089 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.075505] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e59e50-4a5f-4efc-a6f5-a9d5800f67dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.356571] env[65107]: DEBUG nova.compute.utils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1099.365046] env[65107]: INFO nova.compute.manager [None req-178acd92-9c98-43fb-afe4-fb8b22ce7de6 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance to original state: 'active' [ 1099.485419] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.486226] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1099.489058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.680s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.489235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.489392] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1099.489693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.117s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.491095] env[65107]: INFO nova.compute.claims [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1099.496930] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3345f4a-fde6-42e9-a39d-ed70d458f664 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.505871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca3954b-e2f5-4648-822e-3a3c5633cd4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.524749] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889c4e05-72e5-49fc-a02f-5e0ccbe58f23 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.537717] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff07368-faf7-4916-a844-7de9df050a78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.543253] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103521, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.544377] env[65107]: DEBUG nova.network.neutron [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Successfully updated port: 429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1099.577321] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178901MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1099.577493] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1099.814325] env[65107]: DEBUG nova.network.neutron [-] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1099.863374] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.001603] env[65107]: DEBUG nova.compute.utils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1100.003223] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1100.003454] env[65107]: DEBUG nova.network.neutron [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1100.004122] env[65107]: WARNING neutronclient.v2_0.client [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.004122] env[65107]: WARNING neutronclient.v2_0.client [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.004704] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.005060] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.013592] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1100.032059] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103521, 'name': CreateVM_Task, 'duration_secs': 0.596244} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.032388] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1100.032822] env[65107]: WARNING neutronclient.v2_0.client [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.033084] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.033240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.033575] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1100.033964] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4da6ef2-1e26-446b-befc-8c0cf01a0c05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.040300] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1100.040300] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52580ae6-f9eb-dc2f-8976-29176a80ccb8" [ 1100.040300] env[65107]: _type = "Task" [ 1100.040300] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.049524] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.049604] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.049769] env[65107]: DEBUG nova.network.neutron [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1100.051349] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52580ae6-f9eb-dc2f-8976-29176a80ccb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.055853] env[65107]: DEBUG nova.policy [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b03a2312b7d643dd85fceac3d0226ea1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d2bec7b22c74eaa8cbc9c85ebea1e68', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1100.317153] env[65107]: INFO nova.compute.manager [-] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Took 1.31 seconds to deallocate network for instance. [ 1100.364542] env[65107]: DEBUG nova.network.neutron [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Successfully created port: 1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1100.555635] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.555832] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.563275] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52580ae6-f9eb-dc2f-8976-29176a80ccb8, 'name': SearchDatastore_Task, 'duration_secs': 0.010674} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.567018] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1100.567481] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1100.567762] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.567884] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.568138] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1100.569196] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84035315-dff9-4288-a95e-bf952bc54253 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.581168] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1100.581516] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1100.585109] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4d91a23-fe8a-4930-8c0e-08492d796a40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.592565] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1100.592565] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520027e7-29db-eeab-8afb-1f3af75c66ac" [ 1100.592565] env[65107]: _type = "Task" [ 1100.592565] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.601605] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520027e7-29db-eeab-8afb-1f3af75c66ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.630652] env[65107]: DEBUG nova.network.neutron [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1100.670357] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.670495] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.792917] env[65107]: WARNING neutronclient.v2_0.client [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.793677] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.794064] env[65107]: WARNING openstack [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.827707] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.835764] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cff035c-bace-4699-9019-687c0178560e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.844421] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f11c4a-dc95-451d-94b5-a3baf713e856 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.882907] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c103f3d0-0463-4731-8060-600238671067 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.891364] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0552b35e-0842-4926-accb-7565a391b34d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.905844] env[65107]: DEBUG nova.compute.provider_tree [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.925113] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.925380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1100.925649] env[65107]: INFO nova.compute.manager [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Attaching volume c635e8d2-396c-48a5-9b1c-3a8e658ec8be to /dev/sdb [ 1100.948335] env[65107]: DEBUG nova.network.neutron [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updating instance_info_cache with network_info: [{"id": "429e11fc-b198-49e7-bce8-d832c0a6d038", "address": "fa:16:3e:c4:c8:f0", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429e11fc-b1", "ovs_interfaceid": "429e11fc-b198-49e7-bce8-d832c0a6d038", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1100.960358] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945d0bae-7df9-4730-b7f8-2e7cbefbec59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.968938] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b978f314-b00e-49c3-a11e-3b074112e7a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.985607] env[65107]: DEBUG nova.virt.block_device [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updating existing volume attachment record: 0b8ca4a0-d89b-4c22-811d-9bb4aa572d66 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1100.992671] env[65107]: DEBUG nova.compute.manager [req-131727e7-8e62-4787-802b-00a9fe36c302 req-45214b4f-bc5b-4ce8-8fe5-23b37bbcf36d service nova] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Received event network-vif-deleted-8c74ade8-3d3b-4c53-ae8e-85d8e4177ced {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1101.022669] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1101.034488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.034488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.048079] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1101.048337] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1101.048515] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1101.048735] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1101.048925] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1101.049052] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1101.049605] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.049840] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1101.049991] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1101.050397] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1101.050581] env[65107]: DEBUG nova.virt.hardware [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1101.051877] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d38828f-85c8-4d8d-9d91-429a01af2a97 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.061520] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab3d14d-7f91-4115-99a4-22908b1e6297 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.070125] env[65107]: DEBUG nova.compute.manager [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1101.070346] env[65107]: DEBUG nova.compute.manager [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing instance network info cache due to event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1101.070922] env[65107]: DEBUG oslo_concurrency.lockutils [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.071093] env[65107]: DEBUG oslo_concurrency.lockutils [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.071264] env[65107]: DEBUG nova.network.neutron [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1101.081809] env[65107]: WARNING neutronclient.v2_0.client [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.082487] env[65107]: WARNING openstack [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.082850] env[65107]: WARNING openstack [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.105505] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520027e7-29db-eeab-8afb-1f3af75c66ac, 'name': SearchDatastore_Task, 'duration_secs': 0.011242} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.106311] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12b6f940-68d0-45d4-80ab-13046d6b3998 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.113125] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1101.113125] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528dff91-7313-da3b-ceaf-7542d55049e5" [ 1101.113125] env[65107]: _type = "Task" [ 1101.113125] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.124345] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528dff91-7313-da3b-ceaf-7542d55049e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.204342] env[65107]: WARNING openstack [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.204821] env[65107]: WARNING openstack [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.339106] env[65107]: WARNING neutronclient.v2_0.client [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.339900] env[65107]: WARNING openstack [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.340324] env[65107]: WARNING openstack [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.413239] env[65107]: DEBUG nova.scheduler.client.report [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1101.450978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.451413] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance network_info: |[{"id": "429e11fc-b198-49e7-bce8-d832c0a6d038", "address": "fa:16:3e:c4:c8:f0", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429e11fc-b1", "ovs_interfaceid": "429e11fc-b198-49e7-bce8-d832c0a6d038", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1101.453845] env[65107]: DEBUG nova.network.neutron [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updated VIF entry in instance network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1101.454199] env[65107]: DEBUG nova.network.neutron [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1101.455610] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:c8:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f9ffee1-f413-4f28-8bc4-3fb2cf299789', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '429e11fc-b198-49e7-bce8-d832c0a6d038', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1101.463287] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1101.464551] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1101.464806] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0873e41-93c1-41a1-abad-23ba5fcd6ffa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.487095] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1101.487095] env[65107]: value = "task-5103523" [ 1101.487095] env[65107]: _type = "Task" [ 1101.487095] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.496447] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103523, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.537850] env[65107]: INFO nova.compute.manager [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Detaching volume b9b6229b-f202-42ba-8d20-88f1554891f7 [ 1101.581443] env[65107]: INFO nova.virt.block_device [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Attempting to driver detach volume b9b6229b-f202-42ba-8d20-88f1554891f7 from mountpoint /dev/sdb [ 1101.581868] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1101.582221] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992829', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'name': 'volume-b9b6229b-f202-42ba-8d20-88f1554891f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'd2e7eee7-c9b2-49aa-92f9-1efd7c901509', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'serial': 'b9b6229b-f202-42ba-8d20-88f1554891f7'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1101.583333] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a279ea-368d-4622-9387-3529a6ea33fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.608806] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d14abf-ff60-4be9-8659-c988edabd1d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.619830] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145ccdd6-856c-4311-96b8-bf5cc740dfbb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.627503] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528dff91-7313-da3b-ceaf-7542d55049e5, 'name': SearchDatastore_Task, 'duration_secs': 0.010887} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.647747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.648015] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0599df85-ad33-4bd5-b234-611db1ef73f0/0599df85-ad33-4bd5-b234-611db1ef73f0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1101.648501] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40267e14-8c20-41c1-ba5f-e127aadedf1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.651770] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04cf620-6806-4afe-b23a-2d9e9832b5bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.671754] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The volume has not been displaced from its original location: [datastore1] volume-b9b6229b-f202-42ba-8d20-88f1554891f7/volume-b9b6229b-f202-42ba-8d20-88f1554891f7.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1101.677176] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.678774] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3dd282e-7d0d-4bab-ad48-6955eb59b495 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.691724] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1101.691724] env[65107]: value = "task-5103524" [ 1101.691724] env[65107]: _type = "Task" [ 1101.691724] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.698187] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1101.698187] env[65107]: value = "task-5103525" [ 1101.698187] env[65107]: _type = "Task" [ 1101.698187] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.711474] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.918901] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.919786] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1101.924893] env[65107]: DEBUG nova.network.neutron [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Successfully updated port: 1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1101.926096] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.585s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.964925] env[65107]: DEBUG oslo_concurrency.lockutils [req-8358950b-055e-45e1-90d9-a48ca56f5cad req-3e24754f-f312-4d20-bad5-b24fd11b7e96 service nova] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.999444] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103523, 'name': CreateVM_Task, 'duration_secs': 0.368006} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.999641] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1102.000208] env[65107]: WARNING neutronclient.v2_0.client [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.000599] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.000756] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.001103] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1102.001377] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4db93922-3281-4cfd-b878-ec9294028ea9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.008246] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1102.008246] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521e3d9d-f17f-14f4-0813-52da5442aa8b" [ 1102.008246] env[65107]: _type = "Task" [ 1102.008246] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.018112] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521e3d9d-f17f-14f4-0813-52da5442aa8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.203299] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103524, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.212948] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103525, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.429406] env[65107]: DEBUG nova.compute.utils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1102.431058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.431193] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.431361] env[65107]: DEBUG nova.network.neutron [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1102.432499] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1102.432690] env[65107]: DEBUG nova.network.neutron [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1102.433011] env[65107]: WARNING neutronclient.v2_0.client [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.433327] env[65107]: WARNING neutronclient.v2_0.client [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.433901] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.434258] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.443370] env[65107]: INFO nova.compute.claims [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1102.448119] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.448507] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.497641] env[65107]: DEBUG nova.policy [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b73837038b2c4e678c0a7112f6a61f38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd28d5a8ab6f747628edf3aee83355feb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1102.515281] env[65107]: DEBUG nova.network.neutron [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1102.522943] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521e3d9d-f17f-14f4-0813-52da5442aa8b, 'name': SearchDatastore_Task, 'duration_secs': 0.059775} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.523289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.523785] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1102.524075] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.524343] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.524581] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1102.524916] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e7023bb-3da0-4d55-909d-037539ade7a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.535407] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.535809] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.543789] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1102.543985] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1102.544751] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc96b54b-4af3-4599-bc01-563672c9afe4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.551168] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1102.551168] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ec12fd-a978-8cc0-e4a2-46ee83f9c986" [ 1102.551168] env[65107]: _type = "Task" [ 1102.551168] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.559681] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ec12fd-a978-8cc0-e4a2-46ee83f9c986, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.591630] env[65107]: WARNING neutronclient.v2_0.client [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.592360] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.592700] env[65107]: WARNING openstack [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.625987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.625987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.625987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.625987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.625987] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.628151] env[65107]: INFO nova.compute.manager [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Terminating instance [ 1102.675350] env[65107]: DEBUG nova.network.neutron [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updating instance_info_cache with network_info: [{"id": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "address": "fa:16:3e:80:bb:53", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1299e0ad-03", "ovs_interfaceid": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1102.701920] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103524, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562345} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.702215] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0599df85-ad33-4bd5-b234-611db1ef73f0/0599df85-ad33-4bd5-b234-611db1ef73f0.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1102.702431] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1102.702703] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52817644-4356-465f-b0a0-25ebbc0b3f2b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.714096] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103525, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.715508] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1102.715508] env[65107]: value = "task-5103526" [ 1102.715508] env[65107]: _type = "Task" [ 1102.715508] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.724337] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103526, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.804632] env[65107]: DEBUG nova.network.neutron [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Successfully created port: c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1102.934602] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1102.958938] env[65107]: INFO nova.compute.resource_tracker [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating resource usage from migration ff1368a2-a005-471b-bf02-0955e17aa28f [ 1103.066805] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ec12fd-a978-8cc0-e4a2-46ee83f9c986, 'name': SearchDatastore_Task, 'duration_secs': 0.009894} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.067798] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a6698c1-3226-418b-841a-3726865a893b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.076249] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1103.076249] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f95559-3a11-20df-bc6e-692d482194e9" [ 1103.076249] env[65107]: _type = "Task" [ 1103.076249] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.087864] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f95559-3a11-20df-bc6e-692d482194e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.134444] env[65107]: DEBUG nova.compute.manager [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1103.134637] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.135423] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39574daa-a24d-4629-8fa2-58231b98e4db {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.149204] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1103.149204] env[65107]: value = "task-5103527" [ 1103.149204] env[65107]: _type = "Task" [ 1103.149204] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.160348] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.180714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.181128] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Instance network_info: |[{"id": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "address": "fa:16:3e:80:bb:53", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1299e0ad-03", "ovs_interfaceid": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1103.181908] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:bb:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '09bf081b-cdf0-4977-abe2-2339a87409ab', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.196439] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1103.201142] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.202457] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44f07b0e-3d81-4eca-bb12-2ef7ddcbacf4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.239768] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103525, 'name': ReconfigVM_Task, 'duration_secs': 1.29224} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.247488] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1103.252295] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103526, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063049} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.252552] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.252552] env[65107]: value = "task-5103529" [ 1103.252552] env[65107]: _type = "Task" [ 1103.252552] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.253408] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2b07179-3c1d-42b7-bd9f-05adf09005e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.266179] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1103.270012] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa82d01-73d6-4315-84f3-dad45fba53d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.283020] env[65107]: DEBUG nova.compute.manager [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Received event network-vif-plugged-429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1103.283371] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Acquiring lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1103.283663] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1103.283857] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.284087] env[65107]: DEBUG nova.compute.manager [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] No waiting events found dispatching network-vif-plugged-429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1103.284334] env[65107]: WARNING nova.compute.manager [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Received unexpected event network-vif-plugged-429e11fc-b198-49e7-bce8-d832c0a6d038 for instance with vm_state building and task_state spawning. [ 1103.284527] env[65107]: DEBUG nova.compute.manager [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Received event network-changed-429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1103.284686] env[65107]: DEBUG nova.compute.manager [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Refreshing instance network info cache due to event network-changed-429e11fc-b198-49e7-bce8-d832c0a6d038. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1103.284887] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Acquiring lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.285211] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Acquired lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.285249] env[65107]: DEBUG nova.network.neutron [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Refreshing network info cache for port 429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1103.314178] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 0599df85-ad33-4bd5-b234-611db1ef73f0/0599df85-ad33-4bd5-b234-611db1ef73f0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1103.314714] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1103.314714] env[65107]: value = "task-5103530" [ 1103.314714] env[65107]: _type = "Task" [ 1103.314714] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.314968] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103529, 'name': CreateVM_Task} progress is 15%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.320074] env[65107]: WARNING neutronclient.v2_0.client [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.321151] env[65107]: WARNING openstack [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.322027] env[65107]: WARNING openstack [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.330557] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce2f1e5b-49cd-48d2-b3a8-97f7c90ad360 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.360280] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103530, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.362086] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1103.362086] env[65107]: value = "task-5103531" [ 1103.362086] env[65107]: _type = "Task" [ 1103.362086] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.377461] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.422928] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362399e1-c3ce-4fe8-8d4a-f820d415e51a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.433929] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b88441c-ce46-4d44-b9ae-0080c947d9f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.445026] env[65107]: DEBUG nova.compute.manager [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Received event network-vif-plugged-1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1103.445026] env[65107]: DEBUG oslo_concurrency.lockutils [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Acquiring lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1103.445026] env[65107]: DEBUG oslo_concurrency.lockutils [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1103.445026] env[65107]: DEBUG oslo_concurrency.lockutils [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.445026] env[65107]: DEBUG nova.compute.manager [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] No waiting events found dispatching network-vif-plugged-1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1103.445026] env[65107]: WARNING nova.compute.manager [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Received unexpected event network-vif-plugged-1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 for instance with vm_state building and task_state spawning. [ 1103.445026] env[65107]: DEBUG nova.compute.manager [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Received event network-changed-1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1103.445026] env[65107]: DEBUG nova.compute.manager [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Refreshing instance network info cache due to event network-changed-1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1103.445026] env[65107]: DEBUG oslo_concurrency.lockutils [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Acquiring lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.445026] env[65107]: DEBUG oslo_concurrency.lockutils [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Acquired lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.445026] env[65107]: DEBUG nova.network.neutron [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Refreshing network info cache for port 1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1103.481202] env[65107]: WARNING openstack [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.481632] env[65107]: WARNING openstack [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.490622] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34086911-4e85-4566-bead-5e0b2cfab308 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.504394] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f66091-9982-4b29-b288-504850ef001c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.521246] env[65107]: DEBUG nova.compute.provider_tree [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.586682] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f95559-3a11-20df-bc6e-692d482194e9, 'name': SearchDatastore_Task, 'duration_secs': 0.050414} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.586966] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.587250] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1103.587515] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-002357dc-5f89-4133-b9c3-3caa874f2709 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.594291] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1103.594291] env[65107]: value = "task-5103532" [ 1103.594291] env[65107]: _type = "Task" [ 1103.594291] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.604922] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.610470] env[65107]: WARNING neutronclient.v2_0.client [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.611178] env[65107]: WARNING openstack [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.611526] env[65107]: WARNING openstack [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.660182] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103527, 'name': PowerOffVM_Task, 'duration_secs': 0.36601} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.665330] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.665777] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1103.666072] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992848', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'name': 'volume-be6f9921-262e-4dcb-bc60-481f0ad36a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '8cddc4d1-c94c-4f2b-9c4b-3125e4638050', 'attached_at': '2025-12-12T13:29:30.000000', 'detached_at': '', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'serial': 'be6f9921-262e-4dcb-bc60-481f0ad36a02'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1103.667918] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25101a7b-ff57-4c6b-9306-f51dab522eb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.689419] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edf888b-d322-4ded-b6d7-4f8028458859 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.698311] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6343a84-f551-434d-93cc-52d4d4d85992 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.724367] env[65107]: DEBUG nova.network.neutron [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updated VIF entry in instance network info cache for port 429e11fc-b198-49e7-bce8-d832c0a6d038. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1103.724718] env[65107]: DEBUG nova.network.neutron [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updating instance_info_cache with network_info: [{"id": "429e11fc-b198-49e7-bce8-d832c0a6d038", "address": "fa:16:3e:c4:c8:f0", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429e11fc-b1", "ovs_interfaceid": "429e11fc-b198-49e7-bce8-d832c0a6d038", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.726520] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c7bb7d-9322-4aad-8783-7e381dee6635 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.743070] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The volume has not been displaced from its original location: [datastore1] volume-be6f9921-262e-4dcb-bc60-481f0ad36a02/volume-be6f9921-262e-4dcb-bc60-481f0ad36a02.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1103.748353] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfiguring VM instance instance-00000059 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1103.749670] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efa89faa-2de1-4db9-bfbd-695851bc6488 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.772684] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1103.772684] env[65107]: value = "task-5103533" [ 1103.772684] env[65107]: _type = "Task" [ 1103.772684] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.783561] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103529, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.798106] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.857812] env[65107]: DEBUG oslo_vmware.api [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103530, 'name': ReconfigVM_Task, 'duration_secs': 0.171417} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.858147] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992829', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'name': 'volume-b9b6229b-f202-42ba-8d20-88f1554891f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'd2e7eee7-c9b2-49aa-92f9-1efd7c901509', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b6229b-f202-42ba-8d20-88f1554891f7', 'serial': 'b9b6229b-f202-42ba-8d20-88f1554891f7'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1103.873652] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103531, 'name': ReconfigVM_Task, 'duration_secs': 0.314318} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.873949] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 0599df85-ad33-4bd5-b234-611db1ef73f0/0599df85-ad33-4bd5-b234-611db1ef73f0.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1103.874625] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a98872b3-fb89-45aa-9ec2-e3aa48e78432 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.882014] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1103.882014] env[65107]: value = "task-5103534" [ 1103.882014] env[65107]: _type = "Task" [ 1103.882014] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.892030] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103534, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.948212] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1103.953408] env[65107]: WARNING neutronclient.v2_0.client [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.954231] env[65107]: WARNING openstack [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.954567] env[65107]: WARNING openstack [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.979753] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1103.980120] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1103.980371] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1103.980547] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1103.980745] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1103.980920] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1103.981238] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.981371] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1103.981566] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1103.981736] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1103.981904] env[65107]: DEBUG nova.virt.hardware [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1103.982844] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be47170-6dd9-439e-9514-8eb9a6b1bf70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.997817] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a437a9d-fa67-4764-be87-ba3af39addf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.025070] env[65107]: DEBUG nova.scheduler.client.report [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1104.107464] env[65107]: WARNING openstack [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.107893] env[65107]: WARNING openstack [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.122151] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103532, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.167415] env[65107]: WARNING neutronclient.v2_0.client [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.168268] env[65107]: WARNING openstack [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.168730] env[65107]: WARNING openstack [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.230450] env[65107]: DEBUG oslo_concurrency.lockutils [req-f0236139-95fc-4c33-85ed-a4c5b63b7167 req-9db8c1cb-ac09-46ca-85ac-c1cd431e342b service nova] Releasing lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.280371] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103529, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.286531] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103533, 'name': ReconfigVM_Task, 'duration_secs': 0.23154} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.286907] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Reconfigured VM instance instance-00000059 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1104.293122] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bb6a28d-1987-4abf-ad47-c692c1f3fb18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.310174] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1104.310174] env[65107]: value = "task-5103535" [ 1104.310174] env[65107]: _type = "Task" [ 1104.310174] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.324338] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103535, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.327177] env[65107]: DEBUG nova.network.neutron [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updated VIF entry in instance network info cache for port 1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1104.328077] env[65107]: DEBUG nova.network.neutron [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updating instance_info_cache with network_info: [{"id": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "address": "fa:16:3e:80:bb:53", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1299e0ad-03", "ovs_interfaceid": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.394052] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103534, 'name': Rename_Task, 'duration_secs': 0.1753} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.394371] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1104.394618] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4b74f67-4c20-4110-ba84-1289c37ab56f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.402746] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1104.402746] env[65107]: value = "task-5103536" [ 1104.402746] env[65107]: _type = "Task" [ 1104.402746] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.413721] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.415012] env[65107]: DEBUG nova.objects.instance [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'flavor' on Instance uuid d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.530377] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.604s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.530622] env[65107]: INFO nova.compute.manager [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Migrating [ 1104.540023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.198s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.541743] env[65107]: INFO nova.compute.claims [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1104.608724] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647446} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.609048] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1104.609304] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1104.609583] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7f0ff5e-15da-4425-b4b5-be6a873d9a04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.618063] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1104.618063] env[65107]: value = "task-5103537" [ 1104.618063] env[65107]: _type = "Task" [ 1104.618063] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.627479] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103537, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.777561] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103529, 'name': CreateVM_Task, 'duration_secs': 1.185244} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.777758] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.778259] env[65107]: WARNING neutronclient.v2_0.client [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.778629] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.778773] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1104.779175] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1104.779525] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9af8996-5906-4422-a4e0-ff0bb2c11f4d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.784826] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1104.784826] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524ab37c-de92-0d8c-c377-04247414b5f3" [ 1104.784826] env[65107]: _type = "Task" [ 1104.784826] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.792646] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524ab37c-de92-0d8c-c377-04247414b5f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.822840] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103535, 'name': ReconfigVM_Task, 'duration_secs': 0.167552} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.822840] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992848', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'name': 'volume-be6f9921-262e-4dcb-bc60-481f0ad36a02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '8cddc4d1-c94c-4f2b-9c4b-3125e4638050', 'attached_at': '2025-12-12T13:29:30.000000', 'detached_at': '', 'volume_id': 'be6f9921-262e-4dcb-bc60-481f0ad36a02', 'serial': 'be6f9921-262e-4dcb-bc60-481f0ad36a02'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1104.823373] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1104.823829] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878da4d8-ee99-44f9-9531-45e236d910c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.828364] env[65107]: DEBUG nova.network.neutron [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Successfully updated port: c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1104.830504] env[65107]: DEBUG oslo_concurrency.lockutils [req-6fb9ce03-9d6d-4002-93c1-83527c53a99c req-41b1cb02-2352-40ca-9cc8-5756166f60ce service nova] Releasing lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.833993] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1104.834393] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4f98da8-383f-4642-9f33-d60d9b05f2de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.904917] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1104.905219] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1104.905458] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleting the datastore file [datastore1] 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.910266] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a98bfbc8-868c-4593-ae6f-33d5bb61403d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.922452] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103536, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.924923] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1104.924923] env[65107]: value = "task-5103539" [ 1104.924923] env[65107]: _type = "Task" [ 1104.924923] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.934021] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.053786] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.053959] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1105.054219] env[65107]: DEBUG nova.network.neutron [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1105.129382] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103537, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067365} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.129846] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1105.130665] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26b1998-2343-4afb-bb1a-1624289c20f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.156692] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.157043] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cef2073-3b29-4f4d-ab2d-6942671efdd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.179757] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1105.179757] env[65107]: value = "task-5103540" [ 1105.179757] env[65107]: _type = "Task" [ 1105.179757] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.188427] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103540, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.295018] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524ab37c-de92-0d8c-c377-04247414b5f3, 'name': SearchDatastore_Task, 'duration_secs': 0.014576} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.295471] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1105.295639] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1105.295872] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.296023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1105.296207] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1105.296501] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d22a729-60ee-401c-a8c5-8f5bc9248ad2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.305167] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1105.305417] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1105.306235] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18d87b08-0bad-4dfa-a2d1-589eb1dc33fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.311692] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1105.311692] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de9966-cf7f-7723-df3c-7f94e87934c8" [ 1105.311692] env[65107]: _type = "Task" [ 1105.311692] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.320452] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de9966-cf7f-7723-df3c-7f94e87934c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.334873] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "refresh_cache-c2bbaa63-dc53-432d-bcaf-23e698f1f02a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.335072] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "refresh_cache-c2bbaa63-dc53-432d-bcaf-23e698f1f02a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1105.335342] env[65107]: DEBUG nova.network.neutron [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1105.415384] env[65107]: DEBUG oslo_vmware.api [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103536, 'name': PowerOnVM_Task, 'duration_secs': 0.66197} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.415733] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1105.415918] env[65107]: INFO nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Took 8.82 seconds to spawn the instance on the hypervisor. [ 1105.416106] env[65107]: DEBUG nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1105.416951] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5874ccdc-612b-43cb-b1c8-7dd5a59f5824 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.427467] env[65107]: DEBUG oslo_concurrency.lockutils [None req-37ea669b-6209-4fbc-8f4e-e437b49e5014 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.394s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.439034] env[65107]: DEBUG oslo_vmware.api [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220773} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.439178] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.439261] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.439463] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.439658] env[65107]: INFO nova.compute.manager [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Took 2.31 seconds to destroy the instance on the hypervisor. [ 1105.439946] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1105.440183] env[65107]: DEBUG nova.compute.manager [-] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1105.440523] env[65107]: DEBUG nova.network.neutron [-] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1105.440595] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.441173] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.441669] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.493867] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.534382] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1105.534623] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992863', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'name': 'volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04a508b4-4430-411c-8573-726065b558a1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'serial': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1105.535753] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f7f25d-689b-47d1-be29-1f8b429c9fb2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.556894] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946d059a-d558-4479-baac-7c412b9a2fc1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.559583] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.560198] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.560559] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.593669] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be/volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.596887] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62202044-b22a-44aa-9e3d-fbd0849c8f31 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.618250] env[65107]: DEBUG oslo_vmware.api [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1105.618250] env[65107]: value = "task-5103541" [ 1105.618250] env[65107]: _type = "Task" [ 1105.618250] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.631495] env[65107]: DEBUG oslo_vmware.api [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103541, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.694122] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.807665] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.808461] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.831195] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52de9966-cf7f-7723-df3c-7f94e87934c8, 'name': SearchDatastore_Task, 'duration_secs': 0.019101} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.832440] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7c167a3-de29-4137-8cf6-0d81d7cba1ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.839854] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.840272] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.848498] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1105.848498] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d6b93f-012b-ddf5-8ecf-a0d3432d527a" [ 1105.848498] env[65107]: _type = "Task" [ 1105.848498] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.861903] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d6b93f-012b-ddf5-8ecf-a0d3432d527a, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.862356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.862506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.862900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.863159] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.863340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.864985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1105.865265] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/93cd2f0d-6bea-42c2-83b9-085dc94e27ab.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.868729] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00a90e48-95af-4f49-898b-aa092b345326 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.871364] env[65107]: INFO nova.compute.manager [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Terminating instance [ 1105.883589] env[65107]: DEBUG nova.compute.manager [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Received event network-vif-plugged-c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1105.883830] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Acquiring lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.884053] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.884481] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.884481] env[65107]: DEBUG nova.compute.manager [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] No waiting events found dispatching network-vif-plugged-c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1105.884775] env[65107]: WARNING nova.compute.manager [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Received unexpected event network-vif-plugged-c1980f33-8ad2-41e3-96b5-8ea42f4c3399 for instance with vm_state building and task_state spawning. [ 1105.884775] env[65107]: DEBUG nova.compute.manager [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Received event network-changed-c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1105.884867] env[65107]: DEBUG nova.compute.manager [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Refreshing instance network info cache due to event network-changed-c1980f33-8ad2-41e3-96b5-8ea42f4c3399. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1105.884972] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Acquiring lock "refresh_cache-c2bbaa63-dc53-432d-bcaf-23e698f1f02a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.888472] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1105.888472] env[65107]: value = "task-5103542" [ 1105.888472] env[65107]: _type = "Task" [ 1105.888472] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.898854] env[65107]: DEBUG nova.network.neutron [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1105.907274] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.929946] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.930721] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.931338] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.949656] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.949958] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.957842] env[65107]: INFO nova.compute.manager [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Took 13.86 seconds to build instance. [ 1105.983746] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ac6316-d085-4183-bc8b-3d699e005e30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.992063] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df35f690-74f4-4ec4-a6c7-1d11c99ace94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.035414] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a305ab9c-88cc-4d42-8b9e-cc6bb9da2183 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.044066] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71eefd6-d05b-414f-8fcd-628810122060 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.063758] env[65107]: DEBUG nova.compute.provider_tree [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.070828] env[65107]: WARNING neutronclient.v2_0.client [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1106.071810] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.072349] env[65107]: WARNING openstack [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.088271] env[65107]: DEBUG nova.network.neutron [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [{"id": "38203ad5-ce82-479b-8bb4-774b4fb04634", "address": "fa:16:3e:0c:be:68", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38203ad5-ce", "ovs_interfaceid": "38203ad5-ce82-479b-8bb4-774b4fb04634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.129958] env[65107]: DEBUG oslo_vmware.api [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103541, 'name': ReconfigVM_Task, 'duration_secs': 0.406039} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.135795] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be/volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1106.141356] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f17b4ef-6d41-420f-8d16-31a7ceb91542 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.157978] env[65107]: DEBUG oslo_vmware.api [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1106.157978] env[65107]: value = "task-5103543" [ 1106.157978] env[65107]: _type = "Task" [ 1106.157978] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.168923] env[65107]: DEBUG oslo_vmware.api [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103543, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.195058] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103540, 'name': ReconfigVM_Task, 'duration_secs': 0.681222} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.195467] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Reconfigured VM instance instance-00000069 to attach disk [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1106.196185] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baa9c2ee-018e-4718-bdae-33954c7a4eeb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.201695] env[65107]: DEBUG nova.network.neutron [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Updating instance_info_cache with network_info: [{"id": "c1980f33-8ad2-41e3-96b5-8ea42f4c3399", "address": "fa:16:3e:9b:62:04", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1980f33-8a", "ovs_interfaceid": "c1980f33-8ad2-41e3-96b5-8ea42f4c3399", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.204645] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1106.204645] env[65107]: value = "task-5103544" [ 1106.204645] env[65107]: _type = "Task" [ 1106.204645] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.216476] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103544, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.290325] env[65107]: DEBUG nova.compute.manager [req-ef196b2e-eef7-480f-b25c-d303388f49dd req-c20f594b-1028-4d52-81fa-70cec957e5f9 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Received event network-vif-deleted-416a2590-6a42-4a32-a922-a9d5f95786d6 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1106.291333] env[65107]: INFO nova.compute.manager [req-ef196b2e-eef7-480f-b25c-d303388f49dd req-c20f594b-1028-4d52-81fa-70cec957e5f9 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Neutron deleted interface 416a2590-6a42-4a32-a922-a9d5f95786d6; detaching it from the instance and deleting it from the info cache [ 1106.291333] env[65107]: DEBUG nova.network.neutron [req-ef196b2e-eef7-480f-b25c-d303388f49dd req-c20f594b-1028-4d52-81fa-70cec957e5f9 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.375755] env[65107]: DEBUG nova.compute.manager [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1106.376256] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1106.377585] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba9d175-9ca3-46d2-bdbc-491c04ce2757 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.388334] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.388661] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4efa33c8-fafd-433c-9a2b-da1fe14ab8ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.399758] env[65107]: DEBUG oslo_vmware.api [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1106.399758] env[65107]: value = "task-5103545" [ 1106.399758] env[65107]: _type = "Task" [ 1106.399758] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.404229] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103542, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.417970] env[65107]: DEBUG oslo_vmware.api [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103545, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.460265] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e01f9363-4ca8-445f-a723-a4d4bb05e55d tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.379s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1106.463948] env[65107]: DEBUG nova.network.neutron [-] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.568574] env[65107]: DEBUG nova.scheduler.client.report [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1106.591559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1106.670596] env[65107]: DEBUG oslo_vmware.api [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103543, 'name': ReconfigVM_Task, 'duration_secs': 0.257791} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.671071] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992863', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'name': 'volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04a508b4-4430-411c-8573-726065b558a1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'serial': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1106.706012] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "refresh_cache-c2bbaa63-dc53-432d-bcaf-23e698f1f02a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1106.706460] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Instance network_info: |[{"id": "c1980f33-8ad2-41e3-96b5-8ea42f4c3399", "address": "fa:16:3e:9b:62:04", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1980f33-8a", "ovs_interfaceid": "c1980f33-8ad2-41e3-96b5-8ea42f4c3399", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1106.706943] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Acquired lock "refresh_cache-c2bbaa63-dc53-432d-bcaf-23e698f1f02a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.707091] env[65107]: DEBUG nova.network.neutron [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Refreshing network info cache for port c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1106.708293] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:62:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c1980f33-8ad2-41e3-96b5-8ea42f4c3399', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1106.717838] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Creating folder: Project (d28d5a8ab6f747628edf3aee83355feb). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1106.724856] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58d887ca-be08-4cdc-8a69-409a89b25ed8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.733971] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103544, 'name': Rename_Task, 'duration_secs': 0.253846} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.733971] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1106.733971] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82e81643-7b7d-40c8-a5a5-ee4af6eef71b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.738231] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Created folder: Project (d28d5a8ab6f747628edf3aee83355feb) in parent group-v992574. [ 1106.738231] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Creating folder: Instances. Parent ref: group-v992866. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1106.738536] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c497c1fe-b87b-4e74-9c22-bf29370231f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.743885] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1106.743885] env[65107]: value = "task-5103547" [ 1106.743885] env[65107]: _type = "Task" [ 1106.743885] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.751997] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Created folder: Instances in parent group-v992866. [ 1106.751997] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1106.755187] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1106.755508] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103547, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.755729] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68791f86-3c56-45ee-9e36-1d15451e89fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.777239] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1106.777239] env[65107]: value = "task-5103549" [ 1106.777239] env[65107]: _type = "Task" [ 1106.777239] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.786161] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103549, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.794474] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da911a64-906e-46e0-ac73-64d343153ce4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.804377] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac61431a-1f04-451c-bd21-c166174cd884 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.845897] env[65107]: DEBUG nova.compute.manager [req-ef196b2e-eef7-480f-b25c-d303388f49dd req-c20f594b-1028-4d52-81fa-70cec957e5f9 service nova] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Detach interface failed, port_id=416a2590-6a42-4a32-a922-a9d5f95786d6, reason: Instance 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1106.901953] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542597} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.903382] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/93cd2f0d-6bea-42c2-83b9-085dc94e27ab.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.903382] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1106.903382] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3efce01-993b-4481-9195-e2602a9fd750 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.914065] env[65107]: DEBUG oslo_vmware.api [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103545, 'name': PowerOffVM_Task, 'duration_secs': 0.405943} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.915566] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.915789] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1106.916150] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1106.916150] env[65107]: value = "task-5103550" [ 1106.916150] env[65107]: _type = "Task" [ 1106.916150] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.916430] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf8c56ec-5b79-4848-ab9b-9946c9a1ecbb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.928785] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103550, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.967891] env[65107]: INFO nova.compute.manager [-] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Took 1.53 seconds to deallocate network for instance. [ 1107.023737] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1107.023944] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1107.024170] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleting the datastore file [datastore1] d2e7eee7-c9b2-49aa-92f9-1efd7c901509 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1107.024464] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b55422be-3eae-4c19-9af6-153edd1e0fc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.031388] env[65107]: DEBUG oslo_vmware.api [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1107.031388] env[65107]: value = "task-5103552" [ 1107.031388] env[65107]: _type = "Task" [ 1107.031388] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.040720] env[65107]: DEBUG oslo_vmware.api [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103552, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.073353] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.073950] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1107.077459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.693s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1107.225378] env[65107]: WARNING neutronclient.v2_0.client [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.226613] env[65107]: WARNING openstack [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.226959] env[65107]: WARNING openstack [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.256338] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103547, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.289967] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103549, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.348554] env[65107]: WARNING openstack [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.348937] env[65107]: WARNING openstack [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.429186] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103550, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067136} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.429609] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1107.430483] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1292d1b1-47a5-43ab-b5dc-a2dbd06117a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.455793] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/93cd2f0d-6bea-42c2-83b9-085dc94e27ab.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.457218] env[65107]: WARNING neutronclient.v2_0.client [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.457907] env[65107]: WARNING openstack [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.458325] env[65107]: WARNING openstack [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.468037] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e7d5eac-59f5-44e9-9ee4-1e8fe1ee81c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.490971] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1107.490971] env[65107]: value = "task-5103553" [ 1107.490971] env[65107]: _type = "Task" [ 1107.490971] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.500816] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.517092] env[65107]: INFO nova.compute.manager [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Took 0.55 seconds to detach 1 volumes for instance. [ 1107.543018] env[65107]: DEBUG oslo_vmware.api [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103552, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152906} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.543340] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.543681] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.543754] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.543913] env[65107]: INFO nova.compute.manager [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1107.544188] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1107.544442] env[65107]: DEBUG nova.compute.manager [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1107.544560] env[65107]: DEBUG nova.network.neutron [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1107.544826] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.546578] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.546901] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.580635] env[65107]: DEBUG nova.compute.utils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1107.585324] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1107.585710] env[65107]: DEBUG nova.network.neutron [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1107.586334] env[65107]: WARNING neutronclient.v2_0.client [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.586668] env[65107]: WARNING neutronclient.v2_0.client [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.587652] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.588036] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.607469] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.613012] env[65107]: DEBUG nova.network.neutron [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Updated VIF entry in instance network info cache for port c1980f33-8ad2-41e3-96b5-8ea42f4c3399. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1107.613386] env[65107]: DEBUG nova.network.neutron [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Updating instance_info_cache with network_info: [{"id": "c1980f33-8ad2-41e3-96b5-8ea42f4c3399", "address": "fa:16:3e:9b:62:04", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1980f33-8a", "ovs_interfaceid": "c1980f33-8ad2-41e3-96b5-8ea42f4c3399", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1107.669200] env[65107]: DEBUG nova.policy [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d55cff40f12420b8320448083ca8f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7784762e36ae4c6fb30012ec801c3b88', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1107.723277] env[65107]: DEBUG nova.objects.instance [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'flavor' on Instance uuid 04a508b4-4430-411c-8573-726065b558a1 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.758307] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103547, 'name': PowerOnVM_Task} progress is 96%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.791537] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103549, 'name': CreateVM_Task, 'duration_secs': 0.549147} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.791727] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1107.792411] env[65107]: WARNING neutronclient.v2_0.client [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.792999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.793240] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1107.793986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1107.793986] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49906b77-8656-4559-9d33-b0aa2b1b385e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.800792] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1107.800792] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a49f63-be99-ec99-3864-d4c8b10244bb" [ 1107.800792] env[65107]: _type = "Task" [ 1107.800792] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.815410] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a49f63-be99-ec99-3864-d4c8b10244bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.968014] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dff018c-a1f6-406a-9eee-d0ef4af10ad1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.980732] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621c4b55-2e53-4ea8-a24e-e4edf8ec57c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.026513] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.031646] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843f98fe-0431-4d9c-ba3c-c300863e90c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.044773] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65545dd0-f6af-4595-a3ca-542750d6cdf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.051146] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103553, 'name': ReconfigVM_Task, 'duration_secs': 0.422227} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.052149] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/93cd2f0d-6bea-42c2-83b9-085dc94e27ab.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.053014] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6fc6d17-f9ae-4646-96c8-f3ae883b08b6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.067099] env[65107]: DEBUG nova.compute.provider_tree [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.070110] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1108.070110] env[65107]: value = "task-5103554" [ 1108.070110] env[65107]: _type = "Task" [ 1108.070110] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.081237] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103554, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.085415] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1108.108142] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8defa207-8219-4977-889e-04b41ca87f10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.128591] env[65107]: DEBUG oslo_concurrency.lockutils [req-e2966e2e-7031-45db-982a-1ffec2129d42 req-5d449705-95f3-429e-8167-ec5da9a2ff19 service nova] Releasing lock "refresh_cache-c2bbaa63-dc53-432d-bcaf-23e698f1f02a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.128927] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1108.228347] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4022003d-aabe-45ea-9cd0-f85b46411d3a tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.303s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.258287] env[65107]: DEBUG oslo_vmware.api [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103547, 'name': PowerOnVM_Task, 'duration_secs': 1.100906} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.258373] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.258542] env[65107]: INFO nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Took 9.22 seconds to spawn the instance on the hypervisor. [ 1108.259039] env[65107]: DEBUG nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1108.260330] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a3578d-4c48-43f0-bae3-d012b9a8afa4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.315222] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a49f63-be99-ec99-3864-d4c8b10244bb, 'name': SearchDatastore_Task, 'duration_secs': 0.015332} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.316402] env[65107]: DEBUG nova.network.neutron [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Successfully created port: 40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1108.318885] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.319175] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1108.319534] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.319685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.319872] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1108.320405] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29c80d8c-7343-490a-8d8e-93a1c0cc3ee8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.326910] env[65107]: DEBUG nova.compute.manager [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1108.327128] env[65107]: DEBUG nova.compute.manager [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing instance network info cache due to event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1108.327341] env[65107]: DEBUG oslo_concurrency.lockutils [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.327483] env[65107]: DEBUG oslo_concurrency.lockutils [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.327686] env[65107]: DEBUG nova.network.neutron [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1108.356139] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1108.357048] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1108.357713] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18f63c24-7441-44f0-83ec-48def50fee63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.364240] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1108.364240] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d12c14-2377-bec5-afb0-4808a76cd1b3" [ 1108.364240] env[65107]: _type = "Task" [ 1108.364240] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.373986] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d12c14-2377-bec5-afb0-4808a76cd1b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.572462] env[65107]: DEBUG nova.scheduler.client.report [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1108.585735] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103554, 'name': Rename_Task, 'duration_secs': 0.218435} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.586752] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.586752] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f849b62e-a66e-4aa8-8b73-e1d85a0cbb21 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.597644] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1108.597644] env[65107]: value = "task-5103555" [ 1108.597644] env[65107]: _type = "Task" [ 1108.597644] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.609674] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.636892] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.637255] env[65107]: DEBUG nova.network.neutron [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1108.639043] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c6d9bfb-153a-49b1-b117-2faf17ed9d41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.648443] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1108.648443] env[65107]: value = "task-5103556" [ 1108.648443] env[65107]: _type = "Task" [ 1108.648443] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.659657] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.782417] env[65107]: INFO nova.compute.manager [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Took 15.21 seconds to build instance. [ 1108.831334] env[65107]: WARNING neutronclient.v2_0.client [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.832245] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.832711] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.881039] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d12c14-2377-bec5-afb0-4808a76cd1b3, 'name': SearchDatastore_Task, 'duration_secs': 0.025147} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.885746] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b32078a2-fd23-4475-a8c5-14a258ce286e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.895435] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1108.895435] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525109b8-583a-6e4a-42e2-f1ea7435d3e9" [ 1108.895435] env[65107]: _type = "Task" [ 1108.895435] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.914474] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525109b8-583a-6e4a-42e2-f1ea7435d3e9, 'name': SearchDatastore_Task, 'duration_secs': 0.011083} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.914959] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.915361] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] c2bbaa63-dc53-432d-bcaf-23e698f1f02a/c2bbaa63-dc53-432d-bcaf-23e698f1f02a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1108.915760] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-242ffcf2-375e-4b98-ad3e-0b0207b7a820 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.927349] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1108.927349] env[65107]: value = "task-5103557" [ 1108.927349] env[65107]: _type = "Task" [ 1108.927349] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.941864] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.993078] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.994768] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.100418] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1109.115676] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103555, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.136678] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1109.137121] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1109.137220] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1109.137352] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1109.137639] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1109.137822] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1109.139833] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1109.140053] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1109.140240] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1109.140408] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1109.140582] env[65107]: DEBUG nova.virt.hardware [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1109.141507] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358dd8ce-2d1a-4666-83c9-e58b3c413958 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.144512] env[65107]: INFO nova.compute.manager [-] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Took 1.60 seconds to deallocate network for instance. [ 1109.158283] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43471dc0-04bd-4051-9d3e-ba305e186efb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.169018] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103556, 'name': PowerOffVM_Task, 'duration_secs': 0.283669} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.181800] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.182069] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1109.285196] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d3533f4a-d596-46dd-ba5e-ed1feea9942f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.723s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.345741] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.346041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.440794] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103557, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.593105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.514s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.595095] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.349s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.596659] env[65107]: DEBUG nova.objects.instance [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'resources' on Instance uuid e12cb11f-7d6f-47d6-a31e-bf2654174d38 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.618529] env[65107]: DEBUG oslo_vmware.api [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103555, 'name': PowerOnVM_Task, 'duration_secs': 0.5382} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.621022] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1109.621022] env[65107]: INFO nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Took 8.60 seconds to spawn the instance on the hypervisor. [ 1109.621022] env[65107]: DEBUG nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1109.621022] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589a8d66-a5d5-4f64-aa1c-401ed11c8d70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.656494] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.691025] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1109.691025] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1109.691025] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1109.691025] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1109.691025] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1109.691608] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1109.691985] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1109.692329] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1109.692639] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1109.692942] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1109.693327] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1109.699056] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51d8db26-5dfd-4d52-a71b-5eb34d1cab7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.717938] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1109.717938] env[65107]: value = "task-5103558" [ 1109.717938] env[65107]: _type = "Task" [ 1109.717938] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.729947] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.849997] env[65107]: INFO nova.compute.manager [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Detaching volume c635e8d2-396c-48a5-9b1c-3a8e658ec8be [ 1109.875793] env[65107]: WARNING neutronclient.v2_0.client [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.877554] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.878440] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.894736] env[65107]: DEBUG nova.network.neutron [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Successfully updated port: 40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1109.901218] env[65107]: INFO nova.virt.block_device [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Attempting to driver detach volume c635e8d2-396c-48a5-9b1c-3a8e658ec8be from mountpoint /dev/sdb [ 1109.901585] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1109.903172] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992863', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'name': 'volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04a508b4-4430-411c-8573-726065b558a1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'serial': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1109.905969] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e4414c-890f-4fe1-bda4-0027ad69397a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.940857] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aff96f1-90d9-4e6c-b8d7-92cff68c64a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.950085] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103557, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677256} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.951891] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] c2bbaa63-dc53-432d-bcaf-23e698f1f02a/c2bbaa63-dc53-432d-bcaf-23e698f1f02a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1109.952140] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1109.952998] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eac8b0d9-6a0c-4c4e-907c-5eb0ec13e999 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.955274] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9d9460-e096-4e95-853f-b030f5e3a69b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.984972] env[65107]: DEBUG nova.network.neutron [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updated VIF entry in instance network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1109.985402] env[65107]: DEBUG nova.network.neutron [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.987328] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1466d3-10f3-4ff9-bb9e-303b58cdbf6f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.990805] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1109.990805] env[65107]: value = "task-5103559" [ 1109.990805] env[65107]: _type = "Task" [ 1109.990805] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.007696] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The volume has not been displaced from its original location: [datastore2] volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be/volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1110.014181] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1110.015721] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c932efe-1543-4bc8-884a-90ec3532175b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.032469] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103559, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.039243] env[65107]: DEBUG oslo_vmware.api [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1110.039243] env[65107]: value = "task-5103560" [ 1110.039243] env[65107]: _type = "Task" [ 1110.039243] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.049831] env[65107]: DEBUG oslo_vmware.api [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103560, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.143429] env[65107]: INFO nova.compute.manager [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Took 16.44 seconds to build instance. [ 1110.174753] env[65107]: INFO nova.scheduler.client.report [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocation for migration a9997a68-5b5b-4571-b870-f31a9bee6557 [ 1110.240078] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103558, 'name': ReconfigVM_Task, 'duration_secs': 0.399997} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.240226] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1110.411140] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.411391] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1110.411586] env[65107]: DEBUG nova.network.neutron [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1110.432934] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b214273-2278-4cf2-9e05-1b928467feb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.440627] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ee56f5-8c3a-40ab-9c73-decf70439b85 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.478055] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34cf8c0-2fe6-4b32-a8ad-f3cdf42431b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.486250] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bae1aef-8d01-4e7c-a671-833cff5761c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.492305] env[65107]: DEBUG oslo_concurrency.lockutils [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.492604] env[65107]: DEBUG nova.compute.manager [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1110.492776] env[65107]: DEBUG nova.compute.manager [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing instance network info cache due to event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1110.492983] env[65107]: DEBUG oslo_concurrency.lockutils [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.493141] env[65107]: DEBUG oslo_concurrency.lockutils [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1110.493305] env[65107]: DEBUG nova.network.neutron [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1110.507434] env[65107]: DEBUG nova.compute.provider_tree [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.509723] env[65107]: WARNING neutronclient.v2_0.client [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.510756] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.511152] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.529177] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069428} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.529467] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1110.530458] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a6ab4c-3e92-4da2-9f5e-cff4558fec27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.556936] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] c2bbaa63-dc53-432d-bcaf-23e698f1f02a/c2bbaa63-dc53-432d-bcaf-23e698f1f02a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1110.564143] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46d43445-ab1a-4724-a80f-99aee7e7e8f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.585319] env[65107]: DEBUG oslo_vmware.api [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103560, 'name': ReconfigVM_Task, 'duration_secs': 0.50482} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.586760] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1110.591801] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1110.591801] env[65107]: value = "task-5103561" [ 1110.591801] env[65107]: _type = "Task" [ 1110.591801] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.592098] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f0c6e64-68cc-4e2e-9858-112e41d7bdd4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.612878] env[65107]: DEBUG oslo_vmware.api [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1110.612878] env[65107]: value = "task-5103562" [ 1110.612878] env[65107]: _type = "Task" [ 1110.612878] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.617089] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103561, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.632869] env[65107]: DEBUG oslo_vmware.api [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103562, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.645914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5472a5cc-8f38-44a3-adb8-38c45ecee77d tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.952s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.652657] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.653178] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.662787] env[65107]: DEBUG nova.compute.manager [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1110.662996] env[65107]: DEBUG nova.compute.manager [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing instance network info cache due to event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1110.663274] env[65107]: DEBUG oslo_concurrency.lockutils [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.683087] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 17.743s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.726125] env[65107]: WARNING neutronclient.v2_0.client [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.726955] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.727344] env[65107]: WARNING openstack [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.750582] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1110.750808] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1110.750973] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1110.751185] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1110.751948] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1110.751948] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1110.751948] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1110.751948] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1110.752131] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1110.752171] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1110.752459] env[65107]: DEBUG nova.virt.hardware [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1110.758142] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1110.758505] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-139c6d66-f588-4648-8c63-7e64beb04a6a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.783813] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1110.783813] env[65107]: value = "task-5103563" [ 1110.783813] env[65107]: _type = "Task" [ 1110.783813] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.794361] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103563, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.825077] env[65107]: DEBUG nova.network.neutron [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updated VIF entry in instance network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1110.825400] env[65107]: DEBUG nova.network.neutron [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1110.882479] env[65107]: DEBUG nova.compute.manager [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Received event network-vif-deleted-b6bba8b1-799c-4c2b-bd94-8726094948ee {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1110.882479] env[65107]: DEBUG nova.compute.manager [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Received event network-vif-plugged-40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1110.882836] env[65107]: DEBUG oslo_concurrency.lockutils [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1110.883225] env[65107]: DEBUG oslo_concurrency.lockutils [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Lock "bef91b92-add1-4df9-bc24-dab15ce04338-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.883480] env[65107]: DEBUG oslo_concurrency.lockutils [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Lock "bef91b92-add1-4df9-bc24-dab15ce04338-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.883762] env[65107]: DEBUG nova.compute.manager [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] No waiting events found dispatching network-vif-plugged-40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1110.884086] env[65107]: WARNING nova.compute.manager [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Received unexpected event network-vif-plugged-40c68565-653b-4c27-83db-52fff0de54f0 for instance with vm_state building and task_state spawning. [ 1110.884352] env[65107]: DEBUG nova.compute.manager [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Received event network-changed-40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1110.884566] env[65107]: DEBUG nova.compute.manager [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Refreshing instance network info cache due to event network-changed-40c68565-653b-4c27-83db-52fff0de54f0. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1110.884901] env[65107]: DEBUG oslo_concurrency.lockutils [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Acquiring lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.915607] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.916150] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.010224] env[65107]: DEBUG nova.network.neutron [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1111.021268] env[65107]: DEBUG nova.scheduler.client.report [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.054541] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.055061] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.114060] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103561, 'name': ReconfigVM_Task, 'duration_secs': 0.330821} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.114060] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Reconfigured VM instance instance-0000006b to attach disk [datastore2] c2bbaa63-dc53-432d-bcaf-23e698f1f02a/c2bbaa63-dc53-432d-bcaf-23e698f1f02a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.114223] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52856f2e-968a-41f1-b594-b9d1004a143e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.122806] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1111.122806] env[65107]: value = "task-5103564" [ 1111.122806] env[65107]: _type = "Task" [ 1111.122806] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.130637] env[65107]: DEBUG oslo_vmware.api [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103562, 'name': ReconfigVM_Task, 'duration_secs': 0.18754} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.131319] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992863', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'name': 'volume-c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '04a508b4-4430-411c-8573-726065b558a1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be', 'serial': 'c635e8d2-396c-48a5-9b1c-3a8e658ec8be'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1111.138383] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103564, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.141331] env[65107]: WARNING neutronclient.v2_0.client [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.142239] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.142461] env[65107]: WARNING openstack [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.256704] env[65107]: DEBUG nova.network.neutron [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating instance_info_cache with network_info: [{"id": "40c68565-653b-4c27-83db-52fff0de54f0", "address": "fa:16:3e:e0:32:2e", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40c68565-65", "ovs_interfaceid": "40c68565-653b-4c27-83db-52fff0de54f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1111.297809] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103563, 'name': ReconfigVM_Task, 'duration_secs': 0.277864} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.297972] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.298748] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e775e974-ade7-4617-aae3-a587a3bbe60b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.328400] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691/3f544b3f-703a-49d5-b3e8-16672766f691.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.329653] env[65107]: DEBUG oslo_concurrency.lockutils [req-1163c58a-f3d7-48f3-96dc-1ff00452edd2 req-f2ebf12d-4bcc-4a1a-863a-98b14e505ee7 service nova] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.329653] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34bdd0a4-dde2-4535-bd2a-8d03b33ebfc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.342914] env[65107]: DEBUG oslo_concurrency.lockutils [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.343149] env[65107]: DEBUG nova.network.neutron [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1111.350192] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1111.350192] env[65107]: value = "task-5103565" [ 1111.350192] env[65107]: _type = "Task" [ 1111.350192] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.359978] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103565, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.360414] env[65107]: INFO nova.compute.manager [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Rescuing [ 1111.360597] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.360747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.360950] env[65107]: DEBUG nova.network.neutron [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1111.529324] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.532017] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.954s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.553920] env[65107]: INFO nova.scheduler.client.report [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleted allocations for instance e12cb11f-7d6f-47d6-a31e-bf2654174d38 [ 1111.634746] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103564, 'name': Rename_Task, 'duration_secs': 0.155682} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.635085] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1111.635447] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01138571-0a13-44eb-924d-bb1971fe04ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.642552] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1111.642552] env[65107]: value = "task-5103566" [ 1111.642552] env[65107]: _type = "Task" [ 1111.642552] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.654343] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103566, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.685023] env[65107]: DEBUG nova.objects.instance [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'flavor' on Instance uuid 04a508b4-4430-411c-8573-726065b558a1 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.760022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.760436] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance network_info: |[{"id": "40c68565-653b-4c27-83db-52fff0de54f0", "address": "fa:16:3e:e0:32:2e", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40c68565-65", "ovs_interfaceid": "40c68565-653b-4c27-83db-52fff0de54f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1111.760883] env[65107]: DEBUG oslo_concurrency.lockutils [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Acquired lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.760960] env[65107]: DEBUG nova.network.neutron [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Refreshing network info cache for port 40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1111.762325] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:32:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40c68565-653b-4c27-83db-52fff0de54f0', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1111.771427] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1111.772220] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1111.774878] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b900a743-9ee7-4de1-bb07-6aa7d5d5bc42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.796236] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1111.796236] env[65107]: value = "task-5103567" [ 1111.796236] env[65107]: _type = "Task" [ 1111.796236] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.805523] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103567, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.847938] env[65107]: WARNING neutronclient.v2_0.client [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.848645] env[65107]: WARNING openstack [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.848990] env[65107]: WARNING openstack [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.863727] env[65107]: WARNING neutronclient.v2_0.client [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.864509] env[65107]: WARNING openstack [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.864890] env[65107]: WARNING openstack [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.877265] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103565, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.972664] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "7373fbbe-52c1-4955-8da9-a5563d71391b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.972860] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.973152] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.973402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.973679] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.981480] env[65107]: INFO nova.compute.manager [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Terminating instance [ 1112.022332] env[65107]: WARNING openstack [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.022655] env[65107]: WARNING openstack [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.035217] env[65107]: WARNING openstack [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.035655] env[65107]: WARNING openstack [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.064100] env[65107]: DEBUG oslo_concurrency.lockutils [None req-226d9b72-7250-4c1a-9cab-e60460654271 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "e12cb11f-7d6f-47d6-a31e-bf2654174d38" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.564s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.107651] env[65107]: WARNING neutronclient.v2_0.client [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.108339] env[65107]: WARNING openstack [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.108703] env[65107]: WARNING openstack [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.118428] env[65107]: WARNING neutronclient.v2_0.client [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.119218] env[65107]: WARNING openstack [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.119717] env[65107]: WARNING openstack [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.153739] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103566, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.236310] env[65107]: DEBUG nova.network.neutron [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updated VIF entry in instance network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1112.236648] env[65107]: DEBUG nova.network.neutron [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1112.255477] env[65107]: DEBUG nova.network.neutron [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updating instance_info_cache with network_info: [{"id": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "address": "fa:16:3e:80:bb:53", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1299e0ad-03", "ovs_interfaceid": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1112.272716] env[65107]: WARNING neutronclient.v2_0.client [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.273145] env[65107]: WARNING openstack [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.273582] env[65107]: WARNING openstack [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.308554] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103567, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.371625] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103565, 'name': ReconfigVM_Task, 'duration_secs': 0.54312} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.373338] env[65107]: WARNING openstack [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.373697] env[65107]: WARNING openstack [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.380499] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691/3f544b3f-703a-49d5-b3e8-16672766f691.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.380785] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1112.449396] env[65107]: WARNING neutronclient.v2_0.client [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.450098] env[65107]: WARNING openstack [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.450664] env[65107]: WARNING openstack [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.487187] env[65107]: DEBUG nova.compute.manager [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1112.487439] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1112.488913] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c341999b-59c9-4151-b675-da572abe2a7f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.502113] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1112.502325] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e50d4b4-1313-4eee-b0ec-71409feaba4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.510958] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1112.510958] env[65107]: value = "task-5103569" [ 1112.510958] env[65107]: _type = "Task" [ 1112.510958] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.519725] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.546381] env[65107]: DEBUG nova.network.neutron [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updated VIF entry in instance network info cache for port 40c68565-653b-4c27-83db-52fff0de54f0. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1112.546726] env[65107]: DEBUG nova.network.neutron [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating instance_info_cache with network_info: [{"id": "40c68565-653b-4c27-83db-52fff0de54f0", "address": "fa:16:3e:e0:32:2e", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40c68565-65", "ovs_interfaceid": "40c68565-653b-4c27-83db-52fff0de54f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1112.556035] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Applying migration context for instance 3f544b3f-703a-49d5-b3e8-16672766f691 as it has an incoming, in-progress migration ff1368a2-a005-471b-bf02-0955e17aa28f. Migration status is migrating {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1112.557945] env[65107]: INFO nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating resource usage from migration ff1368a2-a005-471b-bf02-0955e17aa28f [ 1112.578983] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.579177] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance ee50d08a-57fc-4c05-96fe-a11fe2708165 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.579309] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 40e17da0-b459-4619-8e9e-18c3093328c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.579431] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3df65e64-0c45-4707-960e-8f1767e2d011 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.579576] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d2e7eee7-c9b2-49aa-92f9-1efd7c901509 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1112.579696] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 2c45d232-a948-4ad7-80a0-589c2dc91ec1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.579812] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 04a508b4-4430-411c-8573-726065b558a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.579936] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. Skipping heal of allocation because we do not know what to do. [ 1112.580083] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.580211] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 7373fbbe-52c1-4955-8da9-a5563d71391b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.580325] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.580454] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance b0b3e29e-9529-4d6d-b27c-c9d0169c5955 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1112.580803] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 0599df85-ad33-4bd5-b234-611db1ef73f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.580803] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d7b4bc1e-c94e-4654-9345-2a8aa945a896 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.580904] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 93cd2f0d-6bea-42c2-83b9-085dc94e27ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.580990] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance c2bbaa63-dc53-432d-bcaf-23e698f1f02a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.581124] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Migration ff1368a2-a005-471b-bf02-0955e17aa28f is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1112.581238] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3f544b3f-703a-49d5-b3e8-16672766f691 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.581352] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bef91b92-add1-4df9-bc24-dab15ce04338 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1112.581561] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1112.581711] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3712MB phys_disk=100GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '15', 'num_vm_active': '11', 'num_task_None': '11', 'num_os_type_None': '15', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '2', 'io_workload': '4', 'num_proj_bfb7fca6907c4c32a1aa6b8023c39afb': '2', 'num_vm_rescued': '2', 'num_proj_7784762e36ae4c6fb30012ec801c3b88': '2', 'num_proj_645feeb0c8524267bff18c5f37f5f011': '1', 'num_proj_544c649453844164bac53c98b342543f': '1', 'num_proj_4778fe9152224fd29d1f6220a19b5a36': '1', 'num_proj_3d2bec7b22c74eaa8cbc9c85ebea1e68': '2', 'num_proj_152868ee0723449baac10f6a8ceeb738': '2', 'num_task_resize_migrating': '1', 'num_proj_c805fe68d0fd4147b9b4b7fdc9d6a6d0': '1', 'num_task_rescuing': '1', 'num_vm_building': '2', 'num_task_spawning': '2', 'num_proj_d28d5a8ab6f747628edf3aee83355feb': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1112.655659] env[65107]: DEBUG oslo_vmware.api [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103566, 'name': PowerOnVM_Task, 'duration_secs': 0.520223} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.656126] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1112.656440] env[65107]: INFO nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Took 8.71 seconds to spawn the instance on the hypervisor. [ 1112.656599] env[65107]: DEBUG nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1112.660693] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b25ab2-49b8-4865-9e1a-2025966d1b79 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.694490] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8ba6ec9e-035a-473e-8c07-f387c8020b12 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.348s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.740953] env[65107]: DEBUG oslo_concurrency.lockutils [req-5c6c342f-12d8-4559-b777-35a88df741a4 req-ef70c539-291d-4bac-b817-dddcf599bd98 service nova] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1112.745654] env[65107]: DEBUG nova.compute.manager [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1112.745654] env[65107]: DEBUG nova.compute.manager [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing instance network info cache due to event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1112.745654] env[65107]: DEBUG oslo_concurrency.lockutils [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.745887] env[65107]: DEBUG oslo_concurrency.lockutils [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.746162] env[65107]: DEBUG nova.network.neutron [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1112.758986] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1112.814384] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103567, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.866114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.866396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.866610] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "04a508b4-4430-411c-8573-726065b558a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.866796] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.866965] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.869660] env[65107]: INFO nova.compute.manager [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Terminating instance [ 1112.875128] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dee65c-e0ef-46a3-80d6-782cce5aa9e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.883499] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cb72fe-66c6-4e45-b0c6-0c541be77ca6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.888935] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b852ba-39ce-4516-996a-4096c0aeb466 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.935690] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917c53d6-2373-4fe4-8be6-ee099c4e3b40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.938869] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e985d119-eb6a-439f-a390-953248be7368 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.961193] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6380b9-b8a3-4392-87f6-9c283509dd88 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.965661] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1112.979550] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.023264] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.051717] env[65107]: DEBUG oslo_concurrency.lockutils [req-1a1e3467-d87f-4d25-b1a7-a1d1584d6293 req-2a6aa24e-f556-4c64-a626-75107b2ee225 service nova] Releasing lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.186329] env[65107]: INFO nova.compute.manager [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Took 18.83 seconds to build instance. [ 1113.249868] env[65107]: WARNING neutronclient.v2_0.client [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.250750] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.251331] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.315022] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103567, 'name': CreateVM_Task, 'duration_secs': 1.446725} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.315270] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1113.315822] env[65107]: WARNING neutronclient.v2_0.client [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.316281] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.316435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.316811] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1113.317141] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df9add74-c183-4ce7-ae0d-b8f0e8c2fb82 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.323394] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1113.323394] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521bd82e-6e94-df19-fd61-a96b58daf37e" [ 1113.323394] env[65107]: _type = "Task" [ 1113.323394] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.333252] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521bd82e-6e94-df19-fd61-a96b58daf37e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.374185] env[65107]: DEBUG nova.compute.manager [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1113.374541] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.375604] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9992755-c9f0-48c6-82f6-a8fbc7bb9d77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.385111] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.386475] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.386866] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.393494] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8a068e9-1d65-4e54-9b09-1935eacaf0e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.401382] env[65107]: DEBUG oslo_vmware.api [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1113.401382] env[65107]: value = "task-5103570" [ 1113.401382] env[65107]: _type = "Task" [ 1113.401382] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.411377] env[65107]: DEBUG oslo_vmware.api [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.456572] env[65107]: WARNING neutronclient.v2_0.client [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.458601] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.459432] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.471813] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.472244] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.482730] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1113.515314] env[65107]: DEBUG nova.network.neutron [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Port 38203ad5-ce82-479b-8bb4-774b4fb04634 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1113.526965] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103569, 'name': PowerOffVM_Task, 'duration_secs': 0.845058} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.527267] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1113.527497] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.527758] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e505ca0e-0325-4121-8b29-37f2fb731155 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.553738] env[65107]: DEBUG nova.network.neutron [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updated VIF entry in instance network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1113.554134] env[65107]: DEBUG nova.network.neutron [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1113.604942] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1113.605458] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1113.605458] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleting the datastore file [datastore1] 7373fbbe-52c1-4955-8da9-a5563d71391b {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.605720] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fb4511e-9de3-449b-a785-85f32c0a8353 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.612944] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for the task: (returnval){ [ 1113.612944] env[65107]: value = "task-5103572" [ 1113.612944] env[65107]: _type = "Task" [ 1113.612944] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.622516] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.689301] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f3f1f3bb-1fef-4af5-bef4-c273dd8d57d6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.346s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.837020] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521bd82e-6e94-df19-fd61-a96b58daf37e, 'name': SearchDatastore_Task, 'duration_secs': 0.016355} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.837348] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.837583] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1113.837818] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.837961] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.838152] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1113.838421] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3f50155-b922-4e1d-b9ba-6fbdcc97951c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.849203] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1113.849404] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1113.850145] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f6d249-0117-4235-917b-e4cdd26f2190 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.855819] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1113.855819] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b23be4-5d71-c697-cb99-4c03b0318f58" [ 1113.855819] env[65107]: _type = "Task" [ 1113.855819] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.863990] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b23be4-5d71-c697-cb99-4c03b0318f58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.913520] env[65107]: DEBUG oslo_vmware.api [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103570, 'name': PowerOffVM_Task, 'duration_secs': 0.30665} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.914213] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1113.914213] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.914213] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-987e4691-1877-4016-8a85-4b1ebc00cfe6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.989173] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1113.989501] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.458s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.989883] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1113.990171] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1113.990388] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleting the datastore file [datastore1] 04a508b4-4430-411c-8573-726065b558a1 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.990696] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.163s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.990916] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.993358] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.967s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.993585] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.995862] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.340s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.996074] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.997957] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-852f08d7-59cc-49b6-a177-bb3ff8c7080e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.007079] env[65107]: DEBUG oslo_vmware.api [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1114.007079] env[65107]: value = "task-5103574" [ 1114.007079] env[65107]: _type = "Task" [ 1114.007079] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.016354] env[65107]: DEBUG oslo_vmware.api [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.025632] env[65107]: INFO nova.scheduler.client.report [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance b0b3e29e-9529-4d6d-b27c-c9d0169c5955 [ 1114.027542] env[65107]: INFO nova.scheduler.client.report [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted allocations for instance d2e7eee7-c9b2-49aa-92f9-1efd7c901509 [ 1114.038272] env[65107]: INFO nova.scheduler.client.report [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted allocations for instance 8cddc4d1-c94c-4f2b-9c4b-3125e4638050 [ 1114.057765] env[65107]: DEBUG oslo_concurrency.lockutils [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.058080] env[65107]: DEBUG nova.compute.manager [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Received event network-changed-429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1114.058256] env[65107]: DEBUG nova.compute.manager [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Refreshing instance network info cache due to event network-changed-429e11fc-b198-49e7-bce8-d832c0a6d038. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1114.058469] env[65107]: DEBUG oslo_concurrency.lockutils [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Acquiring lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.058609] env[65107]: DEBUG oslo_concurrency.lockutils [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Acquired lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.058884] env[65107]: DEBUG nova.network.neutron [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Refreshing network info cache for port 429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1114.124048] env[65107]: DEBUG oslo_vmware.api [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Task: {'id': task-5103572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.434828} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.125915] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.125915] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1114.125915] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1114.126400] env[65107]: INFO nova.compute.manager [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1114.126695] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1114.126926] env[65107]: DEBUG nova.compute.manager [-] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1114.127033] env[65107]: DEBUG nova.network.neutron [-] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1114.127309] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.127842] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.130018] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.173867] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.308771] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.309115] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c6a5482-8fbe-41e3-b86b-68b4b9de9aff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.317895] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1114.317895] env[65107]: value = "task-5103575" [ 1114.317895] env[65107]: _type = "Task" [ 1114.317895] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.329862] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.370029] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b23be4-5d71-c697-cb99-4c03b0318f58, 'name': SearchDatastore_Task, 'duration_secs': 0.014498} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.371291] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f49de48-ec0a-48a7-bfde-ce5de1e1b05f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.379560] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1114.379560] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e56800-fd08-58f2-780b-259cdfe8728c" [ 1114.379560] env[65107]: _type = "Task" [ 1114.379560] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.390446] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e56800-fd08-58f2-780b-259cdfe8728c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.459740] env[65107]: DEBUG nova.compute.manager [req-ef581b30-955a-407b-b53d-0196783883b2 req-6c992851-e646-461c-a9cf-f627c14cd401 service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Received event network-vif-deleted-9e9a520d-8984-4d4d-934b-f46dbdac5f4d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1114.459952] env[65107]: INFO nova.compute.manager [req-ef581b30-955a-407b-b53d-0196783883b2 req-6c992851-e646-461c-a9cf-f627c14cd401 service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Neutron deleted interface 9e9a520d-8984-4d4d-934b-f46dbdac5f4d; detaching it from the instance and deleting it from the info cache [ 1114.460139] env[65107]: DEBUG nova.network.neutron [req-ef581b30-955a-407b-b53d-0196783883b2 req-6c992851-e646-461c-a9cf-f627c14cd401 service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.518012] env[65107]: DEBUG oslo_vmware.api [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37597} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.518289] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.518476] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1114.518654] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1114.518832] env[65107]: INFO nova.compute.manager [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 04a508b4-4430-411c-8573-726065b558a1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1114.519140] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1114.519379] env[65107]: DEBUG nova.compute.manager [-] [instance: 04a508b4-4430-411c-8573-726065b558a1] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1114.520055] env[65107]: DEBUG nova.network.neutron [-] [instance: 04a508b4-4430-411c-8573-726065b558a1] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1114.520055] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.520249] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.520510] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.543716] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d08969c-4880-41a5-91b5-ef14983ef5f2 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "b0b3e29e-9529-4d6d-b27c-c9d0169c5955" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.202s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.553902] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4b2d6aee-fd95-4b02-bcba-458aa1bacc75 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "d2e7eee7-c9b2-49aa-92f9-1efd7c901509" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.691s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.554950] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f7c73e9-e5c2-4d7a-9043-691de27d3388 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "8cddc4d1-c94c-4f2b-9c4b-3125e4638050" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.930s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.562523] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.565444] env[65107]: WARNING neutronclient.v2_0.client [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.566195] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.566761] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.581590] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.582492] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.582492] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.714400] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.714400] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.793432] env[65107]: WARNING neutronclient.v2_0.client [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.794174] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.794542] env[65107]: WARNING openstack [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.830185] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103575, 'name': PowerOffVM_Task, 'duration_secs': 0.357136} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.830575] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.831444] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649496f2-35f9-49dc-a7a6-6721f2f84716 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.857954] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb3bbcc-1e8a-409d-ace1-f6aa4718c3a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.894840] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e56800-fd08-58f2-780b-259cdfe8728c, 'name': SearchDatastore_Task, 'duration_secs': 0.017081} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.897543] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.897875] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1114.898378] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5819302f-9d7c-4549-a7da-4de0e0642edd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.904280] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.904577] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-527cd3ee-d99e-4863-8117-bc6438433777 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.907766] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1114.907766] env[65107]: value = "task-5103576" [ 1114.907766] env[65107]: _type = "Task" [ 1114.907766] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.915559] env[65107]: DEBUG nova.network.neutron [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updated VIF entry in instance network info cache for port 429e11fc-b198-49e7-bce8-d832c0a6d038. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1114.916091] env[65107]: DEBUG nova.network.neutron [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updating instance_info_cache with network_info: [{"id": "429e11fc-b198-49e7-bce8-d832c0a6d038", "address": "fa:16:3e:c4:c8:f0", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap429e11fc-b1", "ovs_interfaceid": "429e11fc-b198-49e7-bce8-d832c0a6d038", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.917541] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1114.917541] env[65107]: value = "task-5103577" [ 1114.917541] env[65107]: _type = "Task" [ 1114.917541] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.924995] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.931487] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1114.931706] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1114.932116] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.932331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.932542] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.932832] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6d20aa2-133c-4336-9bd4-f70b7f538355 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.939143] env[65107]: DEBUG nova.network.neutron [-] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.943306] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.943511] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1114.944280] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9f81bd6-e861-4f7f-831b-d7d68f24d5d7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.951693] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1114.951693] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529b47ba-a454-896e-ba5d-991265f809a2" [ 1114.951693] env[65107]: _type = "Task" [ 1114.951693] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.961544] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529b47ba-a454-896e-ba5d-991265f809a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.963348] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdc37715-d5d0-4b72-bb7e-a9995c38bf61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.977114] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9439e5-38fe-4383-9ed7-42867fc77ea5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.023315] env[65107]: DEBUG nova.compute.manager [req-ef581b30-955a-407b-b53d-0196783883b2 req-6c992851-e646-461c-a9cf-f627c14cd401 service nova] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Detach interface failed, port_id=9e9a520d-8984-4d4d-934b-f46dbdac5f4d, reason: Instance 7373fbbe-52c1-4955-8da9-a5563d71391b could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1115.124130] env[65107]: DEBUG nova.compute.manager [req-4b517bc1-da08-4895-aa94-4d3fd76a34eb req-722aae99-eb53-4388-af8c-2b22564bf23a service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Received event network-vif-deleted-96111c75-74f7-4496-9470-0972fef27f6f {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1115.124312] env[65107]: INFO nova.compute.manager [req-4b517bc1-da08-4895-aa94-4d3fd76a34eb req-722aae99-eb53-4388-af8c-2b22564bf23a service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Neutron deleted interface 96111c75-74f7-4496-9470-0972fef27f6f; detaching it from the instance and deleting it from the info cache [ 1115.124488] env[65107]: DEBUG nova.network.neutron [req-4b517bc1-da08-4895-aa94-4d3fd76a34eb req-722aae99-eb53-4388-af8c-2b22564bf23a service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.421945] env[65107]: DEBUG oslo_concurrency.lockutils [req-4fab0207-84d2-47f3-ac0d-5f5a02a2c295 req-f335a6fc-c661-4980-bc23-ffffb18f3912 service nova] Releasing lock "refresh_cache-d7b4bc1e-c94e-4654-9345-2a8aa945a896" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.422634] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103576, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.442534] env[65107]: INFO nova.compute.manager [-] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Took 1.31 seconds to deallocate network for instance. [ 1115.465026] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529b47ba-a454-896e-ba5d-991265f809a2, 'name': SearchDatastore_Task, 'duration_secs': 0.010236} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.465026] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8e47c55-b5f0-4af4-a10f-4602d7d18235 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.474269] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1115.474269] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c3b8e-301a-fcc0-95c9-eb6a880b9186" [ 1115.474269] env[65107]: _type = "Task" [ 1115.474269] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.491909] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c3b8e-301a-fcc0-95c9-eb6a880b9186, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.496983] env[65107]: DEBUG nova.compute.manager [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1115.498631] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460dd0b1-b14f-479a-9e42-647dddd6b6cc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.527940] env[65107]: DEBUG nova.network.neutron [-] [instance: 04a508b4-4430-411c-8573-726065b558a1] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.561196] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1115.629452] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5585a9bd-8198-4165-8f46-4dc06d2ab4c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.641094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63351673-5b8c-4391-8a07-58dd808fd800 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.654880] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.655157] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.655339] env[65107]: DEBUG nova.network.neutron [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1115.686553] env[65107]: DEBUG nova.compute.manager [req-4b517bc1-da08-4895-aa94-4d3fd76a34eb req-722aae99-eb53-4388-af8c-2b22564bf23a service nova] [instance: 04a508b4-4430-411c-8573-726065b558a1] Detach interface failed, port_id=96111c75-74f7-4496-9470-0972fef27f6f, reason: Instance 04a508b4-4430-411c-8573-726065b558a1 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1115.919846] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612048} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.920735] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1115.920735] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1115.920735] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3714a2a9-dc31-4f2f-818f-a72b665ead3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.929347] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1115.929347] env[65107]: value = "task-5103578" [ 1115.929347] env[65107]: _type = "Task" [ 1115.929347] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.940903] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103578, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.953543] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.953876] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.954119] env[65107]: DEBUG nova.objects.instance [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lazy-loading 'resources' on Instance uuid 7373fbbe-52c1-4955-8da9-a5563d71391b {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.985596] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525c3b8e-301a-fcc0-95c9-eb6a880b9186, 'name': SearchDatastore_Task, 'duration_secs': 0.027717} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.985596] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.986146] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. {{(pid=65107) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1115.986146] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae467487-a65e-4fb8-85ee-5baf1ece24ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.993468] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1115.993468] env[65107]: value = "task-5103579" [ 1115.993468] env[65107]: _type = "Task" [ 1115.993468] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.003692] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.012968] env[65107]: INFO nova.compute.manager [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] instance snapshotting [ 1116.016045] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95acc52f-0412-4f1c-a888-59c11b606a2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.035771] env[65107]: INFO nova.compute.manager [-] [instance: 04a508b4-4430-411c-8573-726065b558a1] Took 1.52 seconds to deallocate network for instance. [ 1116.038623] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf37615-ae49-4fd5-ac5b-982ea78a3a1e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.159963] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.160974] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.161359] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.380507] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.380507] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.395452] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.396154] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.447769] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103578, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068318} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.448340] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1116.449455] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1eb127-9e04-42de-a11c-88be544a7efa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.482530] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1116.484149] env[65107]: WARNING neutronclient.v2_0.client [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.484880] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.485266] env[65107]: WARNING openstack [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.494283] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e99141e7-3258-42ae-bba5-71591added66 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.529167] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103579, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.533981] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1116.533981] env[65107]: value = "task-5103580" [ 1116.533981] env[65107]: _type = "Task" [ 1116.533981] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.546642] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.546955] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103580, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.555307] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1116.555670] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ed1cac04-91cc-4126-a0a3-2993c37c477d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.566265] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1116.566265] env[65107]: value = "task-5103581" [ 1116.566265] env[65107]: _type = "Task" [ 1116.566265] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.584051] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103581, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.630783] env[65107]: DEBUG nova.network.neutron [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [{"id": "38203ad5-ce82-479b-8bb4-774b4fb04634", "address": "fa:16:3e:0c:be:68", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38203ad5-ce", "ovs_interfaceid": "38203ad5-ce82-479b-8bb4-774b4fb04634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1116.708463] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.708758] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.782762] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1cdcbc-7fbd-4845-9ade-797ac7efe1c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.793957] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e255c347-69a6-4aa5-8d4c-751ed0bd2b06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.836562] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef0f92e-3f32-4966-9a61-11bd22090311 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.845204] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4803418b-ddea-4ce6-8add-704d0bcdfd1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.860450] env[65107]: DEBUG nova.compute.provider_tree [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.902708] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1116.930066] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "40e17da0-b459-4619-8e9e-18c3093328c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.930066] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "40e17da0-b459-4619-8e9e-18c3093328c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.930409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "40e17da0-b459-4619-8e9e-18c3093328c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.930409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "40e17da0-b459-4619-8e9e-18c3093328c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.931563] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "40e17da0-b459-4619-8e9e-18c3093328c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.933193] env[65107]: INFO nova.compute.manager [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Terminating instance [ 1117.021458] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566381} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.021737] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk. [ 1117.022568] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b1b5fd-6fef-4e1f-9602-ab0d11862f04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.049313] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.053277] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82ba83e5-da90-4438-ac55-8dd628f6f6fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.071954] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103580, 'name': ReconfigVM_Task, 'duration_secs': 0.308364} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.076401] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfigured VM instance instance-0000006c to attach disk [datastore1] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1117.077196] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1117.077196] env[65107]: value = "task-5103582" [ 1117.077196] env[65107]: _type = "Task" [ 1117.077196] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.077747] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e799a4e1-ebcb-4d54-8c7a-841dbe9d82bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.088012] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103581, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.094458] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103582, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.095840] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1117.095840] env[65107]: value = "task-5103583" [ 1117.095840] env[65107]: _type = "Task" [ 1117.095840] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.103868] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103583, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.134752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.211576] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1117.363663] env[65107]: DEBUG nova.scheduler.client.report [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1117.425718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.436848] env[65107]: DEBUG nova.compute.manager [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1117.437193] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.438121] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f63d7ff-8b7a-47ba-9cd6-b767cbbf1cd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.447315] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.447756] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01a01947-02b7-4f66-b6aa-9a77906e78b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.455426] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1117.455426] env[65107]: value = "task-5103584" [ 1117.455426] env[65107]: _type = "Task" [ 1117.455426] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.464553] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.484172] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.484462] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.582335] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103581, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.591421] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103582, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.605797] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103583, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.661349] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2a8214-758a-46bc-bb36-4375575ccf68 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.693533] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07d3349-b957-4620-bf9d-d49ca963b19a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.704330] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1117.738221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.870065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.915s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.872584] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.326s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.872845] env[65107]: DEBUG nova.objects.instance [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'resources' on Instance uuid 04a508b4-4430-411c-8573-726065b558a1 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.895821] env[65107]: INFO nova.scheduler.client.report [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Deleted allocations for instance 7373fbbe-52c1-4955-8da9-a5563d71391b [ 1117.967113] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103584, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.987051] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1118.084053] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103581, 'name': CreateSnapshot_Task, 'duration_secs': 1.342684} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.087715] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1118.088457] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acadb06e-6da2-4bc6-baa0-7034e7f842ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.103900] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103582, 'name': ReconfigVM_Task, 'duration_secs': 0.837728} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.107152] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab/e439aeac-2298-462c-be63-8218195135cf-rescue.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.108011] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa82f77-2593-4962-8341-dd1b9c5d8842 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.116579] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103583, 'name': Rename_Task, 'duration_secs': 0.780419} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.133211] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.139186] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93327453-86a3-4fd1-bee9-28ec0158fdf3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.141099] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-728cf87e-128b-4513-a16e-5e466d1ff093 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.157604] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1118.157604] env[65107]: value = "task-5103585" [ 1118.157604] env[65107]: _type = "Task" [ 1118.157604] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.160056] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1118.160056] env[65107]: value = "task-5103586" [ 1118.160056] env[65107]: _type = "Task" [ 1118.160056] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.177922] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.183580] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103586, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.214683] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.215205] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7687b49f-99f9-4f19-a6bd-0925a1f7a6d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.223319] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1118.223319] env[65107]: value = "task-5103587" [ 1118.223319] env[65107]: _type = "Task" [ 1118.223319] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.232704] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103587, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.411926] env[65107]: DEBUG oslo_concurrency.lockutils [None req-29858c9f-9059-408a-a927-a2345db84177 tempest-DeleteServersTestJSON-1773899535 tempest-DeleteServersTestJSON-1773899535-project-member] Lock "7373fbbe-52c1-4955-8da9-a5563d71391b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.439s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1118.468987] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103584, 'name': PowerOffVM_Task, 'duration_secs': 0.529032} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.469443] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.469443] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.469678] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e14a9422-f681-49f4-a711-85767bd71c20 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.515904] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.559411] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.559793] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.559981] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleting the datastore file [datastore2] 40e17da0-b459-4619-8e9e-18c3093328c4 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.563654] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17864623-1e7c-422d-807a-ff6f86e70475 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.571748] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1118.571748] env[65107]: value = "task-5103589" [ 1118.571748] env[65107]: _type = "Task" [ 1118.571748] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.585394] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.610411] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1118.614759] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f1fdfe7c-43c3-415c-b3b7-ed95e03eb457 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.626327] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1118.626327] env[65107]: value = "task-5103590" [ 1118.626327] env[65107]: _type = "Task" [ 1118.626327] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.639920] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103590, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.678330] env[65107]: DEBUG oslo_vmware.api [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103585, 'name': PowerOnVM_Task, 'duration_secs': 0.51793} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.683626] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1118.683626] env[65107]: INFO nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Took 9.58 seconds to spawn the instance on the hypervisor. [ 1118.683626] env[65107]: DEBUG nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1118.683918] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103586, 'name': ReconfigVM_Task, 'duration_secs': 0.197123} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.687964] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f99ecd6-62f7-4e65-b2d1-e5de973ad0d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.691154] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.691858] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d5a3208-a3ce-4e92-b476-71d1630263ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.712138] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1118.712138] env[65107]: value = "task-5103592" [ 1118.712138] env[65107]: _type = "Task" [ 1118.712138] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.732087] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.734247] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f784a10f-3ab9-4479-b01d-90b9ed15fd9c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.746087] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103587, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.749323] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f0a796-d302-40d6-bf23-fa67e78e6c79 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.784852] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8eb1fb-35ff-4f80-acbc-fba803911223 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.795352] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f497dd-efe2-44a4-af3b-0949c0a42363 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.816575] env[65107]: DEBUG nova.compute.provider_tree [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.083291] env[65107]: DEBUG oslo_vmware.api [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299822} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.083634] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.083761] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.083946] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.084147] env[65107]: INFO nova.compute.manager [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1119.084423] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1119.084639] env[65107]: DEBUG nova.compute.manager [-] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1119.084741] env[65107]: DEBUG nova.network.neutron [-] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1119.085013] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.085637] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.085885] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.138626] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103590, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.184977] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.226073] env[65107]: INFO nova.compute.manager [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Took 22.90 seconds to build instance. [ 1119.236555] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103592, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.242979] env[65107]: DEBUG oslo_vmware.api [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103587, 'name': PowerOnVM_Task, 'duration_secs': 0.531041} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.243459] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1119.243731] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f67f22e7-4c4f-4918-96b9-7bdab61b9342 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance '3f544b3f-703a-49d5-b3e8-16672766f691' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1119.320747] env[65107]: DEBUG nova.scheduler.client.report [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1119.503165] env[65107]: DEBUG nova.compute.manager [req-c5fc5986-e4b8-497d-8030-213415e50fd7 req-15f1e411-7131-48a2-bbff-083a8d98d41f service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Received event network-vif-deleted-4659264c-9645-4cc9-a6aa-4d3299ae4c31 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1119.503601] env[65107]: INFO nova.compute.manager [req-c5fc5986-e4b8-497d-8030-213415e50fd7 req-15f1e411-7131-48a2-bbff-083a8d98d41f service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Neutron deleted interface 4659264c-9645-4cc9-a6aa-4d3299ae4c31; detaching it from the instance and deleting it from the info cache [ 1119.503931] env[65107]: DEBUG nova.network.neutron [req-c5fc5986-e4b8-497d-8030-213415e50fd7 req-15f1e411-7131-48a2-bbff-083a8d98d41f service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1119.638714] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103590, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.728896] env[65107]: DEBUG oslo_concurrency.lockutils [None req-e406208c-acfe-4482-890f-356240ae8700 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.417s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.735651] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103592, 'name': PowerOnVM_Task} progress is 79%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.826646] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.830154] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.404s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.832467] env[65107]: INFO nova.compute.claims [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.857834] env[65107]: INFO nova.scheduler.client.report [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted allocations for instance 04a508b4-4430-411c-8573-726065b558a1 [ 1119.969415] env[65107]: DEBUG nova.network.neutron [-] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1120.010029] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06f98505-ae5f-4f0f-abe9-48398c6bd3f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.019928] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d05edd1-b364-4460-8b83-e0418aafe2fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.069176] env[65107]: DEBUG nova.compute.manager [req-c5fc5986-e4b8-497d-8030-213415e50fd7 req-15f1e411-7131-48a2-bbff-083a8d98d41f service nova] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Detach interface failed, port_id=4659264c-9645-4cc9-a6aa-4d3299ae4c31, reason: Instance 40e17da0-b459-4619-8e9e-18c3093328c4 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1120.139245] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103590, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.235026] env[65107]: DEBUG oslo_vmware.api [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103592, 'name': PowerOnVM_Task, 'duration_secs': 1.258313} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.235732] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1120.240721] env[65107]: DEBUG nova.compute.manager [None req-cebbcadd-b60e-450f-9992-92713506ec13 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1120.242150] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a856efe3-018d-4014-8896-29983fd6c5ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.372832] env[65107]: DEBUG oslo_concurrency.lockutils [None req-02cfeda2-3587-4912-aafe-93cc35ac04ff tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "04a508b4-4430-411c-8573-726065b558a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.505s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.474758] env[65107]: INFO nova.compute.manager [-] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Took 1.39 seconds to deallocate network for instance. [ 1120.645524] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103590, 'name': CloneVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.750959] env[65107]: DEBUG nova.compute.manager [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Received event network-changed-40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1120.751182] env[65107]: DEBUG nova.compute.manager [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Refreshing instance network info cache due to event network-changed-40c68565-653b-4c27-83db-52fff0de54f0. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1120.751399] env[65107]: DEBUG oslo_concurrency.lockutils [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Acquiring lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.751540] env[65107]: DEBUG oslo_concurrency.lockutils [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Acquired lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.751731] env[65107]: DEBUG nova.network.neutron [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Refreshing network info cache for port 40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1120.986091] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.146514] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103590, 'name': CloneVM_Task, 'duration_secs': 2.079929} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.146945] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Created linked-clone VM from snapshot [ 1121.148094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e993036a-2c22-40ab-a273-f0ddeb37b8ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.153576] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf81c438-e4db-45e1-89dc-7aa394991de9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.166729] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b200ce6-6a93-4be5-89c5-9d972021467b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.172726] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Uploading image 849fc038-618e-402e-a093-4353a4e64502 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1121.208938] env[65107]: INFO nova.compute.manager [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Unrescuing [ 1121.208938] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.208938] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquired lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.209205] env[65107]: DEBUG nova.network.neutron [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1121.212182] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1121.213439] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305bb288-9790-4fd8-bbb2-f34b59b47775 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.219429] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8580fd80-d7ca-414e-ad98-b179fbdcf988 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.232224] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ea18c4-3174-4722-a295-cf6ec313c596 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.236877] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1121.236877] env[65107]: value = "task-5103593" [ 1121.236877] env[65107]: _type = "Task" [ 1121.236877] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.254733] env[65107]: DEBUG nova.compute.provider_tree [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.257153] env[65107]: WARNING neutronclient.v2_0.client [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.258772] env[65107]: WARNING openstack [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.259361] env[65107]: WARNING openstack [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.277861] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103593, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.444844] env[65107]: WARNING openstack [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.445324] env[65107]: WARNING openstack [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.514018] env[65107]: WARNING neutronclient.v2_0.client [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.515215] env[65107]: WARNING openstack [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.515846] env[65107]: WARNING openstack [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.600355] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "3f544b3f-703a-49d5-b3e8-16672766f691" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.600789] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.601047] env[65107]: DEBUG nova.compute.manager [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Going to confirm migration 7 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1121.615637] env[65107]: DEBUG nova.network.neutron [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updated VIF entry in instance network info cache for port 40c68565-653b-4c27-83db-52fff0de54f0. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1121.616105] env[65107]: DEBUG nova.network.neutron [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating instance_info_cache with network_info: [{"id": "40c68565-653b-4c27-83db-52fff0de54f0", "address": "fa:16:3e:e0:32:2e", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40c68565-65", "ovs_interfaceid": "40c68565-653b-4c27-83db-52fff0de54f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1121.713641] env[65107]: WARNING neutronclient.v2_0.client [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.714411] env[65107]: WARNING openstack [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.714749] env[65107]: WARNING openstack [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.752148] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103593, 'name': Destroy_Task, 'duration_secs': 0.402825} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.756541] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Destroyed the VM [ 1121.757106] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1121.758820] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-12272516-6521-4e15-9a72-4ed68bd3b4bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.768684] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1121.768684] env[65107]: value = "task-5103594" [ 1121.768684] env[65107]: _type = "Task" [ 1121.768684] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.769558] env[65107]: DEBUG nova.scheduler.client.report [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1121.784956] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103594, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.824715] env[65107]: WARNING openstack [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.825321] env[65107]: WARNING openstack [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.878153] env[65107]: WARNING neutronclient.v2_0.client [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.878827] env[65107]: WARNING openstack [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.879240] env[65107]: WARNING openstack [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.003014] env[65107]: DEBUG nova.network.neutron [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updating instance_info_cache with network_info: [{"id": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "address": "fa:16:3e:80:bb:53", "network": {"id": "d267f963-f107-412d-9806-f17da1093ac2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-315671503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3d2bec7b22c74eaa8cbc9c85ebea1e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "09bf081b-cdf0-4977-abe2-2339a87409ab", "external-id": "nsx-vlan-transportzone-378", "segmentation_id": 378, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1299e0ad-03", "ovs_interfaceid": "1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1122.109051] env[65107]: WARNING neutronclient.v2_0.client [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.118589] env[65107]: DEBUG oslo_concurrency.lockutils [req-161d0b07-af32-4cce-aac6-3c57a77fcce5 req-be17a5fe-44b3-48ee-8169-b1099db0ab61 service nova] Releasing lock "refresh_cache-bef91b92-add1-4df9-bc24-dab15ce04338" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.170813] env[65107]: WARNING neutronclient.v2_0.client [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.171206] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.171360] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1122.171570] env[65107]: DEBUG nova.network.neutron [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1122.171835] env[65107]: DEBUG nova.objects.instance [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'info_cache' on Instance uuid 3f544b3f-703a-49d5-b3e8-16672766f691 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.278767] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.279452] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1122.283307] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.545s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.285677] env[65107]: INFO nova.compute.claims [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1122.297257] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103594, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.510196] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Releasing lock "refresh_cache-93cd2f0d-6bea-42c2-83b9-085dc94e27ab" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.510713] env[65107]: DEBUG nova.objects.instance [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lazy-loading 'flavor' on Instance uuid 93cd2f0d-6bea-42c2-83b9-085dc94e27ab {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.784738] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103594, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.795617] env[65107]: DEBUG nova.compute.utils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1122.797152] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1122.797364] env[65107]: DEBUG nova.network.neutron [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1122.797691] env[65107]: WARNING neutronclient.v2_0.client [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.798015] env[65107]: WARNING neutronclient.v2_0.client [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.798658] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.799029] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.855669] env[65107]: DEBUG nova.policy [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1122.988799] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.988799] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.021290] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3205533-8837-4c8c-9958-953bcbe66af0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.049029] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1123.049491] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dee09a4-d5bb-4dd4-bb4b-c5d042171642 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.058715] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1123.058715] env[65107]: value = "task-5103595" [ 1123.058715] env[65107]: _type = "Task" [ 1123.058715] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.069189] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.179286] env[65107]: WARNING neutronclient.v2_0.client [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1123.180077] env[65107]: WARNING openstack [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.180497] env[65107]: WARNING openstack [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.203706] env[65107]: DEBUG nova.network.neutron [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Successfully created port: c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1123.288770] env[65107]: DEBUG oslo_vmware.api [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103594, 'name': RemoveSnapshot_Task, 'duration_secs': 1.087587} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.288927] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1123.308379] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1123.491247] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1123.577499] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103595, 'name': PowerOffVM_Task, 'duration_secs': 0.303943} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.577499] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1123.582172] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.587272] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f64c079c-2259-4308-8bed-f94327086779 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.607529] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1123.607529] env[65107]: value = "task-5103596" [ 1123.607529] env[65107]: _type = "Task" [ 1123.607529] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.617711] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103596, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.699533] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7553868c-58af-4040-ba55-83426670afd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.710395] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f4dc93-41bc-401a-90bb-d3004febea17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.750531] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cf866b-449f-41de-9b95-48391bf39ffd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.759409] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5330d2c-ae09-431c-b77f-ee88d0e4141f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.775459] env[65107]: DEBUG nova.compute.provider_tree [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.795207] env[65107]: WARNING nova.compute.manager [None req-a70f5a52-0d34-485d-b584-5829d7c7c15f tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Image not found during snapshot: nova.exception.ImageNotFound: Image 849fc038-618e-402e-a093-4353a4e64502 could not be found. [ 1123.919212] env[65107]: WARNING openstack [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.919611] env[65107]: WARNING openstack [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.016922] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.119835] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103596, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.182580] env[65107]: WARNING neutronclient.v2_0.client [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.183758] env[65107]: WARNING openstack [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.187917] env[65107]: WARNING openstack [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.278838] env[65107]: DEBUG nova.scheduler.client.report [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1124.321328] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1124.358240] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1124.359460] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1124.359460] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1124.359460] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1124.359460] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1124.359460] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1124.359460] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.359780] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1124.359780] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1124.359912] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1124.360209] env[65107]: DEBUG nova.virt.hardware [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1124.361234] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f3578b-3b3c-4915-ace9-f26b73259466 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.371075] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940f0286-37bf-4d4e-9f24-4a5489bb9057 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.425103] env[65107]: DEBUG nova.network.neutron [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [{"id": "38203ad5-ce82-479b-8bb4-774b4fb04634", "address": "fa:16:3e:0c:be:68", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38203ad5-ce", "ovs_interfaceid": "38203ad5-ce82-479b-8bb4-774b4fb04634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.620146] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103596, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.710488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.710801] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.711179] env[65107]: DEBUG nova.objects.instance [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'flavor' on Instance uuid 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.785929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.785929] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1124.789373] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.274s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.790912] env[65107]: INFO nova.compute.claims [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.927754] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-3f544b3f-703a-49d5-b3e8-16672766f691" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.928047] env[65107]: DEBUG nova.objects.instance [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'migration_context' on Instance uuid 3f544b3f-703a-49d5-b3e8-16672766f691 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.002663] env[65107]: DEBUG nova.network.neutron [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Successfully updated port: c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1125.058633] env[65107]: DEBUG nova.compute.manager [req-0b37d635-16f3-4cae-92f3-2597ce8aa0e0 req-3cb2bc30-b041-44f5-a7ec-22be398889c7 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Received event network-vif-plugged-c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1125.058809] env[65107]: DEBUG oslo_concurrency.lockutils [req-0b37d635-16f3-4cae-92f3-2597ce8aa0e0 req-3cb2bc30-b041-44f5-a7ec-22be398889c7 service nova] Acquiring lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.058924] env[65107]: DEBUG oslo_concurrency.lockutils [req-0b37d635-16f3-4cae-92f3-2597ce8aa0e0 req-3cb2bc30-b041-44f5-a7ec-22be398889c7 service nova] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.059107] env[65107]: DEBUG oslo_concurrency.lockutils [req-0b37d635-16f3-4cae-92f3-2597ce8aa0e0 req-3cb2bc30-b041-44f5-a7ec-22be398889c7 service nova] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.059278] env[65107]: DEBUG nova.compute.manager [req-0b37d635-16f3-4cae-92f3-2597ce8aa0e0 req-3cb2bc30-b041-44f5-a7ec-22be398889c7 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] No waiting events found dispatching network-vif-plugged-c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1125.059443] env[65107]: WARNING nova.compute.manager [req-0b37d635-16f3-4cae-92f3-2597ce8aa0e0 req-3cb2bc30-b041-44f5-a7ec-22be398889c7 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Received unexpected event network-vif-plugged-c0845e53-b126-43de-a3da-5356c9c0a9c7 for instance with vm_state building and task_state spawning. [ 1125.119526] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103596, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.125553] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.125784] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.126015] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.126275] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.126457] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.129046] env[65107]: INFO nova.compute.manager [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Terminating instance [ 1125.217255] env[65107]: WARNING neutronclient.v2_0.client [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.218207] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.218632] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.292080] env[65107]: DEBUG nova.compute.utils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1125.294413] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1125.295035] env[65107]: DEBUG nova.network.neutron [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1125.295310] env[65107]: WARNING neutronclient.v2_0.client [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.295858] env[65107]: WARNING neutronclient.v2_0.client [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.296741] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.297267] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.349424] env[65107]: DEBUG nova.policy [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6590cb74e83e45bd944f23fa888a4b04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f379144b78764fe394039d87b043a946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1125.385833] env[65107]: DEBUG nova.objects.instance [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'pci_requests' on Instance uuid 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.430784] env[65107]: DEBUG nova.objects.base [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Object Instance<3f544b3f-703a-49d5-b3e8-16672766f691> lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1125.431950] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4816f61-5408-486a-88a4-a1a076bde3a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.453796] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d9cad63-27ce-4b38-96c2-dee97956b972 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.461073] env[65107]: DEBUG oslo_vmware.api [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1125.461073] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ac4d3e-fa64-768f-e15e-9f86a5b1309b" [ 1125.461073] env[65107]: _type = "Task" [ 1125.461073] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.472136] env[65107]: DEBUG oslo_vmware.api [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ac4d3e-fa64-768f-e15e-9f86a5b1309b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.506157] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-1f3292b7-f6d7-43ca-9caf-8c55794e8311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.506349] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-1f3292b7-f6d7-43ca-9caf-8c55794e8311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.506531] env[65107]: DEBUG nova.network.neutron [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1125.620176] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103596, 'name': ReconfigVM_Task, 'duration_secs': 1.628081} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.620499] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1125.620686] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1125.620947] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88e27817-410f-46fa-a4a3-3cea9cfa7180 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.628502] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1125.628502] env[65107]: value = "task-5103597" [ 1125.628502] env[65107]: _type = "Task" [ 1125.628502] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.632364] env[65107]: DEBUG nova.compute.manager [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1125.632576] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.633751] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fe684e-85f1-49d0-8545-532c16162bc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.642201] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.645766] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7f6c1bd-93e0-48f0-a55d-746ad3c67a41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.647477] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.653803] env[65107]: DEBUG oslo_vmware.api [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1125.653803] env[65107]: value = "task-5103598" [ 1125.653803] env[65107]: _type = "Task" [ 1125.653803] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.664634] env[65107]: DEBUG oslo_vmware.api [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.810573] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1125.827411] env[65107]: DEBUG nova.network.neutron [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Successfully created port: 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1125.888404] env[65107]: DEBUG nova.objects.base [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Object Instance<709eaf30-f1a6-49c2-a0fd-a47eee8aef2d> lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1125.888700] env[65107]: DEBUG nova.network.neutron [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1125.889115] env[65107]: WARNING neutronclient.v2_0.client [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.889684] env[65107]: WARNING neutronclient.v2_0.client [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.890102] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.891459] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.977493] env[65107]: DEBUG oslo_vmware.api [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ac4d3e-fa64-768f-e15e-9f86a5b1309b, 'name': SearchDatastore_Task, 'duration_secs': 0.011614} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.978149] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.004601] env[65107]: DEBUG nova.policy [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1126.009665] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.009665] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.143347] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.149676] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837e4b6b-6685-4d32-a615-d045cacf4f85 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.161258] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4ed9cb-9fc6-4ff9-bf54-cfefb89f6f56 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.168065] env[65107]: DEBUG oslo_vmware.api [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103598, 'name': PowerOffVM_Task, 'duration_secs': 0.263628} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.168796] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.168992] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.169325] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-668cd538-49a7-43fc-b09a-3ee26d6204d1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.200610] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ddf2aa-e494-4740-987d-cbf76d9cb04e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.210440] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004d5ed4-ff53-4fc9-a061-8bcbeac4874c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.226218] env[65107]: DEBUG nova.compute.provider_tree [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.247595] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.247957] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.248241] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleting the datastore file [datastore2] c2bbaa63-dc53-432d-bcaf-23e698f1f02a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.248530] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40697b01-edd0-431e-861e-6085adfb8249 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.257786] env[65107]: DEBUG oslo_vmware.api [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1126.257786] env[65107]: value = "task-5103600" [ 1126.257786] env[65107]: _type = "Task" [ 1126.257786] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.270482] env[65107]: DEBUG oslo_vmware.api [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.641569] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.730024] env[65107]: DEBUG nova.scheduler.client.report [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1126.769756] env[65107]: DEBUG oslo_vmware.api [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221327} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.769989] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.770170] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.770362] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1126.770538] env[65107]: INFO nova.compute.manager [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1126.770807] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1126.771059] env[65107]: DEBUG nova.compute.manager [-] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1126.771167] env[65107]: DEBUG nova.network.neutron [-] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1126.771435] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.771988] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.772259] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.823201] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1126.857107] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1126.857464] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1126.857646] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1126.857831] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1126.857977] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1126.858165] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1126.858403] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.858561] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1126.858726] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1126.858884] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1126.859146] env[65107]: DEBUG nova.virt.hardware [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1126.860025] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f758b8d7-7554-4b68-84c1-feece700ec26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.870388] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb28b44a-e491-435e-82a0-2f4baa2592fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.140634] env[65107]: DEBUG oslo_vmware.api [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103597, 'name': PowerOnVM_Task, 'duration_secs': 1.203485} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.140961] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1127.141218] env[65107]: DEBUG nova.compute.manager [None req-3841ef51-5fd3-4640-89d3-37a318f13da2 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1127.142228] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb7ef4a-cca9-479f-9509-a14701e92348 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.236444] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1127.237115] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1127.240319] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.254s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.240571] env[65107]: DEBUG nova.objects.instance [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'resources' on Instance uuid 40e17da0-b459-4619-8e9e-18c3093328c4 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.640412] env[65107]: DEBUG nova.network.neutron [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Successfully updated port: 12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1127.674169] env[65107]: DEBUG nova.network.neutron [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1127.744436] env[65107]: DEBUG nova.compute.utils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1127.749546] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1127.749546] env[65107]: DEBUG nova.network.neutron [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1127.749996] env[65107]: WARNING neutronclient.v2_0.client [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.750265] env[65107]: WARNING neutronclient.v2_0.client [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.750868] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.751248] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.913672] env[65107]: DEBUG nova.network.neutron [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Successfully updated port: 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1127.977543] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.977950] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.985262] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1128.014036] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8abba0e-d2c8-42c0-b622-56d0ce5aa268 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.020480] env[65107]: DEBUG nova.policy [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a974f6d9394deab176e30972bb7277', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dced4f9aef2f49cf990203b693533aa1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1128.028622] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2a1fd0-4fcc-4a42-a47f-525f33bd2c71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.065450] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc2a28f-18e3-4923-a1d0-54c64fbb5cab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.074232] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea05d182-a3fd-4866-a94e-ee64a1b5fe90 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.089661] env[65107]: DEBUG nova.compute.provider_tree [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.147557] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.147557] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1128.147557] env[65107]: DEBUG nova.network.neutron [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1128.249972] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1128.266941] env[65107]: WARNING neutronclient.v2_0.client [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1128.267671] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1128.268018] env[65107]: WARNING openstack [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1128.419759] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.419964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1128.420166] env[65107]: DEBUG nova.network.neutron [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1128.575244] env[65107]: DEBUG nova.network.neutron [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Successfully created port: 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1128.593545] env[65107]: DEBUG nova.scheduler.client.report [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1128.653579] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1128.653579] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1128.765453] env[65107]: DEBUG nova.network.neutron [-] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1128.923968] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1128.924511] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1128.987124] env[65107]: DEBUG nova.network.neutron [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Updating instance_info_cache with network_info: [{"id": "c0845e53-b126-43de-a3da-5356c9c0a9c7", "address": "fa:16:3e:a6:f2:bd", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0845e53-b1", "ovs_interfaceid": "c0845e53-b126-43de-a3da-5356c9c0a9c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1129.070934] env[65107]: WARNING nova.network.neutron [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] 40a76aa5-ac46-4067-a98d-cd60410bf479 already exists in list: networks containing: ['40a76aa5-ac46-4067-a98d-cd60410bf479']. ignoring it [ 1129.099225] env[65107]: DEBUG nova.network.neutron [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1129.103104] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.106644] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.090s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.108927] env[65107]: INFO nova.compute.claims [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1129.158162] env[65107]: INFO nova.scheduler.client.report [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleted allocations for instance 40e17da0-b459-4619-8e9e-18c3093328c4 [ 1129.261330] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1129.268900] env[65107]: INFO nova.compute.manager [-] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Took 2.50 seconds to deallocate network for instance. [ 1129.298379] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1129.298705] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1129.298871] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1129.299073] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1129.299433] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1129.299968] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1129.299968] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.300131] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1129.300240] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1129.300430] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1129.300575] env[65107]: DEBUG nova.virt.hardware [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1129.301855] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5ab218-1637-46ca-9aba-4165d77d9e06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.312617] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57578a9-11df-448b-9a47-ab727b5d6780 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.490438] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-1f3292b7-f6d7-43ca-9caf-8c55794e8311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1129.490920] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Instance network_info: |[{"id": "c0845e53-b126-43de-a3da-5356c9c0a9c7", "address": "fa:16:3e:a6:f2:bd", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0845e53-b1", "ovs_interfaceid": "c0845e53-b126-43de-a3da-5356c9c0a9c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1129.491416] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:f2:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0845e53-b126-43de-a3da-5356c9c0a9c7', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1129.499177] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1129.499437] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1129.499674] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de72359c-4073-4a19-b812-a2030872c3a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.525861] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1129.525861] env[65107]: value = "task-5103601" [ 1129.525861] env[65107]: _type = "Task" [ 1129.525861] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.535109] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103601, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.668923] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d012a2ba-586c-476f-ac5d-6a3aaba782b1 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "40e17da0-b459-4619-8e9e-18c3093328c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.739s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.778136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.034622] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.035227] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.046894] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103601, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.050961] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.051467] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.206185] env[65107]: WARNING neutronclient.v2_0.client [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.207130] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.207703] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.220948] env[65107]: DEBUG nova.network.neutron [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Successfully updated port: 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1130.271030] env[65107]: WARNING neutronclient.v2_0.client [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.271710] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.272058] env[65107]: WARNING openstack [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.375717] env[65107]: DEBUG nova.network.neutron [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1130.391867] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.392290] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.449418] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6068f9c-ca6f-4a8a-8188-dd50cb0b8242 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.458259] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228dee9d-aa5e-4c9f-afd1-0f858afbf409 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.463416] env[65107]: WARNING neutronclient.v2_0.client [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.464124] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.464449] env[65107]: WARNING openstack [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.497828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08fc042-fa3c-4798-b007-33957b0d6928 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.508147] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff378db7-f4c0-46a9-ac05-15ffc8a0be55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.522829] env[65107]: DEBUG nova.compute.provider_tree [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.544677] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103601, 'name': CreateVM_Task, 'duration_secs': 0.655394} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.544945] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1130.545492] env[65107]: WARNING neutronclient.v2_0.client [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.545769] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.545973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1130.549346] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1130.549346] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe1e270a-78aa-45c0-894d-98a44da9dc76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.555074] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1130.555074] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5209d2a5-f198-6686-6561-b7fdebca07b0" [ 1130.555074] env[65107]: _type = "Task" [ 1130.555074] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.566642] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5209d2a5-f198-6686-6561-b7fdebca07b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.573561] env[65107]: DEBUG nova.compute.manager [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Received event network-changed-c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1130.573933] env[65107]: DEBUG nova.compute.manager [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Refreshing instance network info cache due to event network-changed-c0845e53-b126-43de-a3da-5356c9c0a9c7. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1130.574072] env[65107]: DEBUG oslo_concurrency.lockutils [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Acquiring lock "refresh_cache-1f3292b7-f6d7-43ca-9caf-8c55794e8311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.574114] env[65107]: DEBUG oslo_concurrency.lockutils [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Acquired lock "refresh_cache-1f3292b7-f6d7-43ca-9caf-8c55794e8311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1130.574270] env[65107]: DEBUG nova.network.neutron [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Refreshing network info cache for port c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1130.581814] env[65107]: DEBUG nova.network.neutron [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "address": "fa:16:3e:ec:8b:30", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12711066-1d", "ovs_interfaceid": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1130.595899] env[65107]: DEBUG nova.compute.manager [req-c4e3a99e-f36e-402f-b5cd-caf4b605a53b req-6a335e55-25d2-4f50-9014-8d48cd8f1718 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-vif-plugged-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1130.596713] env[65107]: DEBUG oslo_concurrency.lockutils [req-c4e3a99e-f36e-402f-b5cd-caf4b605a53b req-6a335e55-25d2-4f50-9014-8d48cd8f1718 service nova] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.596941] env[65107]: DEBUG oslo_concurrency.lockutils [req-c4e3a99e-f36e-402f-b5cd-caf4b605a53b req-6a335e55-25d2-4f50-9014-8d48cd8f1718 service nova] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.597471] env[65107]: DEBUG oslo_concurrency.lockutils [req-c4e3a99e-f36e-402f-b5cd-caf4b605a53b req-6a335e55-25d2-4f50-9014-8d48cd8f1718 service nova] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.597471] env[65107]: DEBUG nova.compute.manager [req-c4e3a99e-f36e-402f-b5cd-caf4b605a53b req-6a335e55-25d2-4f50-9014-8d48cd8f1718 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] No waiting events found dispatching network-vif-plugged-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1130.597606] env[65107]: WARNING nova.compute.manager [req-c4e3a99e-f36e-402f-b5cd-caf4b605a53b req-6a335e55-25d2-4f50-9014-8d48cd8f1718 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received unexpected event network-vif-plugged-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 for instance with vm_state active and task_state None. [ 1130.648226] env[65107]: DEBUG nova.compute.manager [req-45b20bae-9bbd-45a2-bc3b-54104e2eba66 req-1e698401-f198-4a7c-bfe9-535d6d14f67e service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-vif-plugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1130.648359] env[65107]: DEBUG oslo_concurrency.lockutils [req-45b20bae-9bbd-45a2-bc3b-54104e2eba66 req-1e698401-f198-4a7c-bfe9-535d6d14f67e service nova] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.648498] env[65107]: DEBUG oslo_concurrency.lockutils [req-45b20bae-9bbd-45a2-bc3b-54104e2eba66 req-1e698401-f198-4a7c-bfe9-535d6d14f67e service nova] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.648666] env[65107]: DEBUG oslo_concurrency.lockutils [req-45b20bae-9bbd-45a2-bc3b-54104e2eba66 req-1e698401-f198-4a7c-bfe9-535d6d14f67e service nova] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.648833] env[65107]: DEBUG nova.compute.manager [req-45b20bae-9bbd-45a2-bc3b-54104e2eba66 req-1e698401-f198-4a7c-bfe9-535d6d14f67e service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] No waiting events found dispatching network-vif-plugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1130.648998] env[65107]: WARNING nova.compute.manager [req-45b20bae-9bbd-45a2-bc3b-54104e2eba66 req-1e698401-f198-4a7c-bfe9-535d6d14f67e service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received unexpected event network-vif-plugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 for instance with vm_state building and task_state spawning. [ 1130.725346] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.725511] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1130.725694] env[65107]: DEBUG nova.network.neutron [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1130.880159] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.880928] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance network_info: |[{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1130.881192] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:b5:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd903c404-a23a-40c0-a217-96d4bb2e5b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b5b909b-b28a-42bf-ad8f-3dd0e7df6811', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.889382] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1130.889678] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1130.889887] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ec54a99-3416-4ca9-91c8-c1b2995e426c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.912948] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.912948] env[65107]: value = "task-5103602" [ 1130.912948] env[65107]: _type = "Task" [ 1130.912948] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.922534] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103602, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.026756] env[65107]: DEBUG nova.scheduler.client.report [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1131.068521] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5209d2a5-f198-6686-6561-b7fdebca07b0, 'name': SearchDatastore_Task, 'duration_secs': 0.040967} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.068784] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.069046] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.069296] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.069440] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.069616] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.069894] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1522b184-b3f8-48da-ab4d-2088cbf547cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.081670] env[65107]: WARNING neutronclient.v2_0.client [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.081791] env[65107]: WARNING openstack [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.082118] env[65107]: WARNING openstack [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.089491] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.089642] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.091879] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.091879] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.091879] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.092076] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7385679-7fd9-4174-aed9-b176f623817c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.096075] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775d2010-747c-4e59-897b-eff02daea431 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.105132] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1131.105132] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52266ad9-9bc2-322f-4552-ab201305b688" [ 1131.105132] env[65107]: _type = "Task" [ 1131.105132] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.117964] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1131.118332] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1131.118495] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1131.118702] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1131.118822] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1131.118963] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1131.119197] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1131.119339] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1131.119527] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1131.119725] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1131.119902] env[65107]: DEBUG nova.virt.hardware [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1131.126570] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Reconfiguring VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1131.130728] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3681ceb-9e0d-4968-aa64-5e00a811b283 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.154038] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52266ad9-9bc2-322f-4552-ab201305b688, 'name': SearchDatastore_Task, 'duration_secs': 0.012279} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.156269] env[65107]: DEBUG oslo_vmware.api [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1131.156269] env[65107]: value = "task-5103603" [ 1131.156269] env[65107]: _type = "Task" [ 1131.156269] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.156535] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2464c5bc-9fea-4b6f-8b9c-ea6b1fc94a37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.170478] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1131.170478] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520dc94d-910e-18a2-75c9-9e83d2f0d5b7" [ 1131.170478] env[65107]: _type = "Task" [ 1131.170478] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.181454] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520dc94d-910e-18a2-75c9-9e83d2f0d5b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.217728] env[65107]: WARNING openstack [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.218290] env[65107]: WARNING openstack [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.230461] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.230917] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.287187] env[65107]: DEBUG nova.network.neutron [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1131.309492] env[65107]: WARNING neutronclient.v2_0.client [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.310247] env[65107]: WARNING openstack [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.311120] env[65107]: WARNING openstack [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.324409] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.325116] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.426613] env[65107]: WARNING neutronclient.v2_0.client [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.427296] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.427781] env[65107]: WARNING openstack [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.440771] env[65107]: DEBUG nova.network.neutron [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Updated VIF entry in instance network info cache for port c0845e53-b126-43de-a3da-5356c9c0a9c7. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1131.441373] env[65107]: DEBUG nova.network.neutron [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Updating instance_info_cache with network_info: [{"id": "c0845e53-b126-43de-a3da-5356c9c0a9c7", "address": "fa:16:3e:a6:f2:bd", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0845e53-b1", "ovs_interfaceid": "c0845e53-b126-43de-a3da-5356c9c0a9c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1131.449600] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103602, 'name': CreateVM_Task, 'duration_secs': 0.391973} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.450026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1131.450324] env[65107]: WARNING neutronclient.v2_0.client [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.450679] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.450836] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.451472] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1131.453057] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc71c3b1-54bc-4f7d-984c-0b93b534e1f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.460660] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1131.460660] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52937177-5aff-90a9-37e3-0f0cf0346aa1" [ 1131.460660] env[65107]: _type = "Task" [ 1131.460660] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.475602] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52937177-5aff-90a9-37e3-0f0cf0346aa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.534654] env[65107]: DEBUG nova.network.neutron [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1131.535957] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.536464] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1131.539314] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.561s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.670137] env[65107]: DEBUG oslo_vmware.api [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103603, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.680741] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520dc94d-910e-18a2-75c9-9e83d2f0d5b7, 'name': SearchDatastore_Task, 'duration_secs': 0.013516} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.681042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.681332] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 1f3292b7-f6d7-43ca-9caf-8c55794e8311/1f3292b7-f6d7-43ca-9caf-8c55794e8311.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1131.681605] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5909f81-0257-49f4-ba75-78578a0c94cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.689681] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1131.689681] env[65107]: value = "task-5103604" [ 1131.689681] env[65107]: _type = "Task" [ 1131.689681] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.698885] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.908283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "ee50d08a-57fc-4c05-96fe-a11fe2708165" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.908634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.908822] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "ee50d08a-57fc-4c05-96fe-a11fe2708165-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.909038] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.909825] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.912987] env[65107]: INFO nova.compute.manager [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Terminating instance [ 1131.936029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.936319] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.944951] env[65107]: DEBUG oslo_concurrency.lockutils [req-19ee0638-32ac-46d0-ab33-515609f30126 req-420ea8a5-a40d-46a9-b448-6521e9d0e265 service nova] Releasing lock "refresh_cache-1f3292b7-f6d7-43ca-9caf-8c55794e8311" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.974097] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52937177-5aff-90a9-37e3-0f0cf0346aa1, 'name': SearchDatastore_Task, 'duration_secs': 0.017965} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.974435] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.974679] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.974975] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.975170] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.975333] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.975629] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0084a67f-d9dc-4ef9-888d-8e36255d3d0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.986153] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.986359] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.987146] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a29b51db-b093-4af3-9d27-6e5dc55feb1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.997843] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1131.997843] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52506555-e6c3-0fa4-6dcc-414110af8fa0" [ 1131.997843] env[65107]: _type = "Task" [ 1131.997843] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.008482] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52506555-e6c3-0fa4-6dcc-414110af8fa0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.014276] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.014786] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.015167] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.015441] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.015673] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.018284] env[65107]: INFO nova.compute.manager [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Terminating instance [ 1132.040768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.040768] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance network_info: |[{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1132.042209] env[65107]: DEBUG nova.compute.utils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1132.043774] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:c2:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72f482ac-0a93-4cf4-9d54-a7e812ca4cd9', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1132.052221] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1132.055449] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1132.055651] env[65107]: DEBUG nova.network.neutron [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1132.056008] env[65107]: WARNING neutronclient.v2_0.client [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.056336] env[65107]: WARNING neutronclient.v2_0.client [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.056895] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.057245] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.064877] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1132.065458] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0be5a8eb-1144-4991-8932-143dc97fb70b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.091211] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.091211] env[65107]: value = "task-5103605" [ 1132.091211] env[65107]: _type = "Task" [ 1132.091211] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.104787] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103605, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.108734] env[65107]: DEBUG nova.policy [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd87f09a43c049deb1fda21602b55a64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '544c649453844164bac53c98b342543f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1132.174902] env[65107]: DEBUG oslo_vmware.api [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103603, 'name': ReconfigVM_Task, 'duration_secs': 0.85631} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.175419] env[65107]: WARNING neutronclient.v2_0.client [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.175506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.175737] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Reconfigured VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1132.203438] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103604, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.419585] env[65107]: DEBUG nova.compute.manager [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1132.419585] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.420649] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a4300c-53d6-495c-846b-a31924f649fc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.432014] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.432563] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6bd1c8e-e0d9-4b53-968e-ec420c03b1b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.438903] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1132.443924] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87796727-27cc-421a-a336-5e5ef639149c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.449429] env[65107]: DEBUG oslo_vmware.api [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1132.449429] env[65107]: value = "task-5103606" [ 1132.449429] env[65107]: _type = "Task" [ 1132.449429] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.458384] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f6dd06-9674-4b37-b483-51956ba9da4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.466693] env[65107]: DEBUG oslo_vmware.api [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.508986] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac2bbba-fa74-436f-92b4-78a849b7fc41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.519477] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52506555-e6c3-0fa4-6dcc-414110af8fa0, 'name': SearchDatastore_Task, 'duration_secs': 0.012725} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.522513] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bebd890-d9bd-4351-8eb9-8410c9d1161a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.526183] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ca3209-b588-4cfe-b05d-34918c883e8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.530899] env[65107]: DEBUG nova.compute.manager [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1132.531119] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.532268] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ea0c11-f851-4b9a-bb78-80f1c2370f03 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.539887] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1132.539887] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee1b6a-794a-ce1b-5fff-aea8fc0b698e" [ 1132.539887] env[65107]: _type = "Task" [ 1132.539887] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.551361] env[65107]: DEBUG nova.compute.provider_tree [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.552868] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.556713] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1132.559391] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfbedccd-5a2c-48f7-b65c-4a768c71731f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.569159] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee1b6a-794a-ce1b-5fff-aea8fc0b698e, 'name': SearchDatastore_Task, 'duration_secs': 0.012623} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.572020] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.572462] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1132.573286] env[65107]: DEBUG oslo_vmware.api [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1132.573286] env[65107]: value = "task-5103607" [ 1132.573286] env[65107]: _type = "Task" [ 1132.573286] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.573611] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16f2218f-a319-4290-b12e-9db0321d8fe7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.581384] env[65107]: DEBUG nova.network.neutron [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Successfully created port: 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1132.591420] env[65107]: DEBUG oslo_vmware.api [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.593945] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1132.593945] env[65107]: value = "task-5103608" [ 1132.593945] env[65107]: _type = "Task" [ 1132.593945] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.607968] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103608, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.612033] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103605, 'name': CreateVM_Task, 'duration_secs': 0.415285} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.612224] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1132.612740] env[65107]: WARNING neutronclient.v2_0.client [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.613136] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.613332] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.613623] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1132.613902] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bcc3a9c-b04e-4765-93b7-8d693125f06c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.620129] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1132.620129] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dfa564-e2b0-010d-2fd6-1c56e2b122b0" [ 1132.620129] env[65107]: _type = "Task" [ 1132.620129] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.631259] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dfa564-e2b0-010d-2fd6-1c56e2b122b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.651725] env[65107]: DEBUG nova.compute.manager [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1132.651725] env[65107]: DEBUG nova.compute.manager [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing instance network info cache due to event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1132.651725] env[65107]: DEBUG oslo_concurrency.lockutils [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.651725] env[65107]: DEBUG oslo_concurrency.lockutils [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.653364] env[65107]: DEBUG nova.network.neutron [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1132.683755] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c475317b-ce0f-4648-a57c-19a774abe746 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.973s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.687836] env[65107]: DEBUG nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-changed-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1132.688157] env[65107]: DEBUG nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing instance network info cache due to event network-changed-12711066-1ddb-4f73-aaf9-f2eb3d8c4515. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1132.688411] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.688558] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.688719] env[65107]: DEBUG nova.network.neutron [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing network info cache for port 12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1132.703212] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.773022} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.703550] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 1f3292b7-f6d7-43ca-9caf-8c55794e8311/1f3292b7-f6d7-43ca-9caf-8c55794e8311.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1132.703810] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1132.704608] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d288f91-3a80-4c9a-a381-98efb483da35 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.715960] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1132.715960] env[65107]: value = "task-5103609" [ 1132.715960] env[65107]: _type = "Task" [ 1132.715960] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.728172] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.970784] env[65107]: DEBUG oslo_vmware.api [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103606, 'name': PowerOffVM_Task, 'duration_secs': 0.273383} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.971301] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1132.971539] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1132.972729] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.973048] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a47cf6dc-4473-4cf3-9eab-60096f7cf3e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.058979] env[65107]: DEBUG nova.scheduler.client.report [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1133.076235] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1133.076525] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1133.076716] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleting the datastore file [datastore2] ee50d08a-57fc-4c05-96fe-a11fe2708165 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.078651] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-090d87b0-2b86-4b70-92e0-3456a27e7672 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.094865] env[65107]: DEBUG oslo_vmware.api [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103607, 'name': PowerOffVM_Task, 'duration_secs': 0.277501} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.096926] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.097229] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1133.098220] env[65107]: DEBUG oslo_vmware.api [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for the task: (returnval){ [ 1133.098220] env[65107]: value = "task-5103611" [ 1133.098220] env[65107]: _type = "Task" [ 1133.098220] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.098955] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6883dc43-47df-418f-a6a2-56d17f91284f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.118960] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103608, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.124047] env[65107]: DEBUG oslo_vmware.api [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.137356] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52dfa564-e2b0-010d-2fd6-1c56e2b122b0, 'name': SearchDatastore_Task, 'duration_secs': 0.024002} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.137766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.138152] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.138442] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.138584] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.138849] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.139115] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3a4c0a4-e805-435a-9b45-c3ef08a3c0f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.155992] env[65107]: WARNING neutronclient.v2_0.client [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.157249] env[65107]: WARNING openstack [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.157697] env[65107]: WARNING openstack [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.171808] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.171808] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1133.172342] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9e4973c-2611-47cb-bee6-4ca9f7d50e4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.180550] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1133.180550] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528c57e2-583b-1270-4dc5-df1f78c324ef" [ 1133.180550] env[65107]: _type = "Task" [ 1133.180550] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.191969] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528c57e2-583b-1270-4dc5-df1f78c324ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.193127] env[65107]: WARNING neutronclient.v2_0.client [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.193868] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.194298] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.210340] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1133.210731] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1133.210817] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Deleting the datastore file [datastore2] 93cd2f0d-6bea-42c2-83b9-085dc94e27ab {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.211044] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1851fa7-b7a7-4dee-9813-3ba4a2ecc33d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.222339] env[65107]: DEBUG oslo_vmware.api [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1133.222339] env[65107]: value = "task-5103613" [ 1133.222339] env[65107]: _type = "Task" [ 1133.222339] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.229489] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.313947} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.230423] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1133.231343] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd0de79-9336-4e7a-9126-e5dc95a6ce20 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.237760] env[65107]: DEBUG oslo_vmware.api [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103613, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.259904] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 1f3292b7-f6d7-43ca-9caf-8c55794e8311/1f3292b7-f6d7-43ca-9caf-8c55794e8311.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.260903] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7785b425-9c8b-452c-bdfa-ca65a0d92fcf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.283357] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1133.283357] env[65107]: value = "task-5103614" [ 1133.283357] env[65107]: _type = "Task" [ 1133.283357] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.295447] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103614, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.522146] env[65107]: WARNING openstack [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.522668] env[65107]: WARNING openstack [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.567489] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.567987] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.580330] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1133.620214] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103608, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635636} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.622828] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1133.623117] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1133.623282] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1133.623455] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1133.623594] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1133.623738] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1133.623992] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.624179] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1133.624348] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1133.624509] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1133.624684] env[65107]: DEBUG nova.virt.hardware [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1133.628435] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.628663] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.629489] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a268b379-e2e5-47fb-b35b-dd0886d6dfbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.632615] env[65107]: DEBUG oslo_vmware.api [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Task: {'id': task-5103611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272048} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.637384] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29367cde-2ad5-4ef6-8bfa-61395d25e862 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.639849] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.640053] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.640749] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.640749] env[65107]: INFO nova.compute.manager [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1133.640749] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1133.641477] env[65107]: DEBUG nova.compute.manager [-] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1133.641575] env[65107]: DEBUG nova.network.neutron [-] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1133.641820] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.642422] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.642682] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.655392] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6f8a81-2177-4848-ac3c-995e48c5360f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.663434] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1133.663434] env[65107]: value = "task-5103615" [ 1133.663434] env[65107]: _type = "Task" [ 1133.663434] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.689957] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103615, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.699186] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528c57e2-583b-1270-4dc5-df1f78c324ef, 'name': SearchDatastore_Task, 'duration_secs': 0.020822} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.699186] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-904a266d-d336-4753-b861-29753f509796 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.705936] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1133.705936] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520df3d4-62d9-d202-2725-33f6d184e1a3" [ 1133.705936] env[65107]: _type = "Task" [ 1133.705936] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.717685] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520df3d4-62d9-d202-2725-33f6d184e1a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.733959] env[65107]: DEBUG oslo_vmware.api [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103613, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190731} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.734266] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.734452] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.734626] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.734799] env[65107]: INFO nova.compute.manager [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1133.735118] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1133.735340] env[65107]: DEBUG nova.compute.manager [-] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1133.735785] env[65107]: DEBUG nova.network.neutron [-] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1133.735785] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.736306] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.736602] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.744552] env[65107]: WARNING neutronclient.v2_0.client [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.745311] env[65107]: WARNING openstack [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.745725] env[65107]: WARNING openstack [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.768126] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.793698] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103614, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.795335] env[65107]: WARNING neutronclient.v2_0.client [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.795989] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.796369] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.852178] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.085115] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.546s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.092088] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.310s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.092088] env[65107]: DEBUG nova.objects.instance [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lazy-loading 'resources' on Instance uuid c2bbaa63-dc53-432d-bcaf-23e698f1f02a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.174614] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103615, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084538} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.174937] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.175787] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7cfe7e-0368-4bf0-85a6-1fcaf5fec16d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.201846] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.202517] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a4108ac-c788-4aa1-88cd-22576b2b1e74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.226654] env[65107]: DEBUG nova.network.neutron [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Successfully updated port: 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1134.234487] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1134.234487] env[65107]: value = "task-5103616" [ 1134.234487] env[65107]: _type = "Task" [ 1134.234487] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.242512] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520df3d4-62d9-d202-2725-33f6d184e1a3, 'name': SearchDatastore_Task, 'duration_secs': 0.022618} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.242703] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1134.243514] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1134.243514] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0922ded-55f9-443c-88d6-e49ba88f5bfe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.250769] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103616, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.257658] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1134.257658] env[65107]: value = "task-5103617" [ 1134.257658] env[65107]: _type = "Task" [ 1134.257658] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.269542] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.295232] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103614, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.307372] env[65107]: DEBUG nova.network.neutron [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updated VIF entry in instance network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1134.307899] env[65107]: DEBUG nova.network.neutron [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.424605] env[65107]: DEBUG nova.network.neutron [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updated VIF entry in instance network info cache for port 12711066-1ddb-4f73-aaf9-f2eb3d8c4515. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1134.425109] env[65107]: DEBUG nova.network.neutron [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "address": "fa:16:3e:ec:8b:30", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12711066-1d", "ovs_interfaceid": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.600438] env[65107]: DEBUG nova.network.neutron [-] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.661028] env[65107]: INFO nova.scheduler.client.report [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocation for migration ff1368a2-a005-471b-bf02-0955e17aa28f [ 1134.694046] env[65107]: DEBUG nova.compute.manager [req-865a7afa-bd74-4584-ac5d-bbc028222e9e req-24084e92-a571-45bd-81c9-69df3067c31f service nova] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Received event network-vif-deleted-d5026810-9268-4f3f-9e44-06ea55e41569 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.694046] env[65107]: DEBUG nova.compute.manager [req-865a7afa-bd74-4584-ac5d-bbc028222e9e req-24084e92-a571-45bd-81c9-69df3067c31f service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Received event network-vif-deleted-1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.694046] env[65107]: INFO nova.compute.manager [req-865a7afa-bd74-4584-ac5d-bbc028222e9e req-24084e92-a571-45bd-81c9-69df3067c31f service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Neutron deleted interface 1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5; detaching it from the instance and deleting it from the info cache [ 1134.694046] env[65107]: DEBUG nova.network.neutron [req-865a7afa-bd74-4584-ac5d-bbc028222e9e req-24084e92-a571-45bd-81c9-69df3067c31f service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.745775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.746021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.746182] env[65107]: DEBUG nova.network.neutron [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1134.748924] env[65107]: DEBUG nova.compute.manager [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Received event network-vif-plugged-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.749521] env[65107]: DEBUG oslo_concurrency.lockutils [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.749521] env[65107]: DEBUG oslo_concurrency.lockutils [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.749663] env[65107]: DEBUG oslo_concurrency.lockutils [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.749840] env[65107]: DEBUG nova.compute.manager [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] No waiting events found dispatching network-vif-plugged-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1134.750861] env[65107]: WARNING nova.compute.manager [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Received unexpected event network-vif-plugged-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be for instance with vm_state building and task_state spawning. [ 1134.750861] env[65107]: DEBUG nova.compute.manager [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Received event network-changed-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.750861] env[65107]: DEBUG nova.compute.manager [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Refreshing instance network info cache due to event network-changed-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1134.750861] env[65107]: DEBUG oslo_concurrency.lockutils [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Acquiring lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.767459] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103616, 'name': ReconfigVM_Task, 'duration_secs': 0.348166} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.771290] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.772993] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49de54a9-7665-402e-bb09-e6b85c02ac0d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.779602] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103617, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.797314] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1134.797314] env[65107]: value = "task-5103618" [ 1134.797314] env[65107]: _type = "Task" [ 1134.797314] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.806435] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103614, 'name': ReconfigVM_Task, 'duration_secs': 1.071161} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.807300] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 1f3292b7-f6d7-43ca-9caf-8c55794e8311/1f3292b7-f6d7-43ca-9caf-8c55794e8311.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.808012] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00d39e6b-bf37-4e2e-91bd-6114594d24f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.816470] env[65107]: DEBUG oslo_concurrency.lockutils [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1134.816801] env[65107]: DEBUG nova.compute.manager [req-8cb9ea9b-9b19-4a10-970f-bfe472dd794c req-d2c231f1-4074-4086-b027-0bedbaba4ecc service nova] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Received event network-vif-deleted-c1980f33-8ad2-41e3-96b5-8ea42f4c3399 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.817633] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103618, 'name': Rename_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.825381] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1134.825381] env[65107]: value = "task-5103619" [ 1134.825381] env[65107]: _type = "Task" [ 1134.825381] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.847035] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103619, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.880629] env[65107]: DEBUG nova.network.neutron [-] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.928077] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1134.928355] env[65107]: DEBUG nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-vif-plugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.928779] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.928845] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.928997] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.929162] env[65107]: DEBUG nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] No waiting events found dispatching network-vif-plugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1134.929352] env[65107]: WARNING nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received unexpected event network-vif-plugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 for instance with vm_state building and task_state spawning. [ 1134.929490] env[65107]: DEBUG nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.929644] env[65107]: DEBUG nova.compute.manager [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing instance network info cache due to event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1134.930098] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.930098] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.930218] env[65107]: DEBUG nova.network.neutron [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1134.933678] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae232504-6227-4bef-8e50-492fd89e7649 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.945448] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07573778-bcb9-4136-86a7-1c8f4b00e451 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.982820] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.983126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.985943] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8b82e2-a0dd-4362-8973-cf0f5c94626c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.995634] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b912ff62-4ba9-42ac-8b89-b0cd13385eb2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.011351] env[65107]: DEBUG nova.compute.provider_tree [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.103755] env[65107]: INFO nova.compute.manager [-] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Took 1.46 seconds to deallocate network for instance. [ 1135.172510] env[65107]: DEBUG oslo_concurrency.lockutils [None req-73dbbffb-6129-4a81-b978-a30a0a069fec tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.572s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.195650] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89e3ee12-4558-4525-bb10-37609901e5f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.208266] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0be66a-6fd6-4c1c-9823-8545b48c91a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.245482] env[65107]: DEBUG nova.compute.manager [req-865a7afa-bd74-4584-ac5d-bbc028222e9e req-24084e92-a571-45bd-81c9-69df3067c31f service nova] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Detach interface failed, port_id=1299e0ad-03f4-4023-bc6d-5fd1c1c1a5f5, reason: Instance 93cd2f0d-6bea-42c2-83b9-085dc94e27ab could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1135.259310] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.259712] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.278238] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571378} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.278517] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1135.278734] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1135.278994] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-835d4aad-a932-422d-b14f-3046a55239ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.286236] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1135.286236] env[65107]: value = "task-5103620" [ 1135.286236] env[65107]: _type = "Task" [ 1135.286236] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.294962] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103620, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.309884] env[65107]: DEBUG nova.network.neutron [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1135.312202] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103618, 'name': Rename_Task, 'duration_secs': 0.181134} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.313043] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.313043] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5264c21-5c0f-4030-b71b-79363c5f0836 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.321160] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1135.321160] env[65107]: value = "task-5103621" [ 1135.321160] env[65107]: _type = "Task" [ 1135.321160] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.333084] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103621, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.334508] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.334913] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.348598] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103619, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.382757] env[65107]: INFO nova.compute.manager [-] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Took 1.65 seconds to deallocate network for instance. [ 1135.414998] env[65107]: WARNING neutronclient.v2_0.client [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.415728] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.416234] env[65107]: WARNING openstack [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.438489] env[65107]: WARNING neutronclient.v2_0.client [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.439305] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.439748] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.486906] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.487126] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.488371] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79ccb0d-5510-4614-9de2-eb0efec34dbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.507818] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08586cfb-3c38-4286-b0f4-6d9892b8170c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.532256] env[65107]: DEBUG nova.scheduler.client.report [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1135.535959] env[65107]: WARNING neutronclient.v2_0.client [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.541873] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Reconfiguring VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1135.544592] env[65107]: DEBUG nova.network.neutron [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updating instance_info_cache with network_info: [{"id": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "address": "fa:16:3e:f6:22:6e", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3914e94e-1c", "ovs_interfaceid": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1135.546138] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd4f4bbe-1fa0-4388-801e-3909497b44fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.573363] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1135.573363] env[65107]: value = "task-5103622" [ 1135.573363] env[65107]: _type = "Task" [ 1135.573363] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.583399] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.601593] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.601979] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.610459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.667318] env[65107]: WARNING neutronclient.v2_0.client [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.668078] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.668482] env[65107]: WARNING openstack [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.753344] env[65107]: DEBUG nova.network.neutron [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updated VIF entry in instance network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1135.755977] env[65107]: DEBUG nova.network.neutron [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1135.797736] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103620, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086478} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.798067] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1135.798907] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e94f29-a619-4f87-9915-3626a10899b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.822399] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.822912] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11e608f9-e29e-4d8d-8b72-176bfb2c57ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.851309] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103621, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.856309] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1135.856309] env[65107]: value = "task-5103623" [ 1135.856309] env[65107]: _type = "Task" [ 1135.856309] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.856713] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103619, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.889722] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.047647] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.050832] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.078s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.052559] env[65107]: INFO nova.compute.claims [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.060893] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.061332] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Instance network_info: |[{"id": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "address": "fa:16:3e:f6:22:6e", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3914e94e-1c", "ovs_interfaceid": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1136.061685] env[65107]: DEBUG oslo_concurrency.lockutils [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Acquired lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.061881] env[65107]: DEBUG nova.network.neutron [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Refreshing network info cache for port 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1136.063117] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:22:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3914e94e-1c0c-410a-adcb-0ac7cb7ff3be', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1136.073066] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1136.073066] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1136.073228] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4077ab1-a51c-44e1-8f18-b57742269d65 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.090568] env[65107]: INFO nova.scheduler.client.report [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleted allocations for instance c2bbaa63-dc53-432d-bcaf-23e698f1f02a [ 1136.112408] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.115496] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1136.115496] env[65107]: value = "task-5103624" [ 1136.115496] env[65107]: _type = "Task" [ 1136.115496] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.970405] env[65107]: DEBUG oslo_concurrency.lockutils [req-effd52e1-1d43-4a71-a611-840d1108495d req-5d602388-bcea-4af2-beac-dc91dbcd044c service nova] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.971110] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.971411] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "3f544b3f-703a-49d5-b3e8-16672766f691" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.971619] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.971819] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.971985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.972152] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.977359] env[65107]: WARNING neutronclient.v2_0.client [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.977359] env[65107]: WARNING openstack [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.977359] env[65107]: WARNING openstack [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.987956] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103624, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.987956] env[65107]: WARNING oslo_vmware.common.loopingcall [-] task run outlasted interval by 0.372445 sec [ 1136.988587] env[65107]: INFO nova.compute.manager [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Terminating instance [ 1136.990226] env[65107]: DEBUG oslo_concurrency.lockutils [None req-2372d89c-4aed-4641-a280-efb3823f711e tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "c2bbaa63-dc53-432d-bcaf-23e698f1f02a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.864s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.000683] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.005916] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.011500] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.013813] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.025067] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.025397] env[65107]: DEBUG oslo_vmware.api [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103621, 'name': PowerOnVM_Task, 'duration_secs': 0.649681} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.025634] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103619, 'name': Rename_Task, 'duration_secs': 1.20311} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.025890] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103623, 'name': ReconfigVM_Task, 'duration_secs': 0.404138} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.030973] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.031169] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1137.031429] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1137.031626] env[65107]: INFO nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Took 10.21 seconds to spawn the instance on the hypervisor. [ 1137.031799] env[65107]: DEBUG nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1137.033336] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1137.033791] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.034680] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.038659] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.038822] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Cleaning up deleted instances with incomplete migration {{(pid=65107) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 1137.040563] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe231c4-90a8-4f5b-82b5-e6c89905c00a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.043681] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ea27b96-cfea-49bd-9b64-caeee4d433f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.046319] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-827fb197-58c1-4cff-b2c0-c4ced183becf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.053782] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103624, 'name': CreateVM_Task, 'duration_secs': 0.386033} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.055125] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1137.055774] env[65107]: WARNING neutronclient.v2_0.client [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.056326] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.056551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1137.057314] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1137.058232] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b06dfaa1-6379-4c3e-b798-ce2a78914b04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.065072] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1137.065072] env[65107]: value = "task-5103625" [ 1137.065072] env[65107]: _type = "Task" [ 1137.065072] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.068548] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1137.068548] env[65107]: value = "task-5103626" [ 1137.068548] env[65107]: _type = "Task" [ 1137.068548] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.078705] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1137.078705] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520d45d6-5edc-c34b-d553-058af8d545d5" [ 1137.078705] env[65107]: _type = "Task" [ 1137.078705] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.090768] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103626, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.091117] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.102238] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520d45d6-5edc-c34b-d553-058af8d545d5, 'name': SearchDatastore_Task, 'duration_secs': 0.01231} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.102238] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.102507] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1137.104023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.104023] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1137.104023] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1137.104023] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2c7b4b8-f28b-4039-a897-e3e9ed2e2d08 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.115440] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1137.115670] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1137.116665] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3969288-4c12-4a05-af66-89f96942b8a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.123664] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1137.123664] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e41fea-c871-7e3b-00c3-e4dd84b64f2c" [ 1137.123664] env[65107]: _type = "Task" [ 1137.123664] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.140540] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e41fea-c871-7e3b-00c3-e4dd84b64f2c, 'name': SearchDatastore_Task, 'duration_secs': 0.01129} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.141597] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7772ab7-6039-474d-84af-b55b0607ef1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.147837] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1137.147837] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523688c2-536d-1464-4c7d-a514c213d54f" [ 1137.147837] env[65107]: _type = "Task" [ 1137.147837] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.153830] env[65107]: WARNING openstack [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.154330] env[65107]: WARNING openstack [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.165106] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523688c2-536d-1464-4c7d-a514c213d54f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.196665] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "0161c1ba-8aff-4f70-a407-caa991e43f15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.196919] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1137.237795] env[65107]: WARNING neutronclient.v2_0.client [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.238511] env[65107]: WARNING openstack [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.238854] env[65107]: WARNING openstack [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.337353] env[65107]: DEBUG nova.network.neutron [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updated VIF entry in instance network info cache for port 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1137.337805] env[65107]: DEBUG nova.network.neutron [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updating instance_info_cache with network_info: [{"id": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "address": "fa:16:3e:f6:22:6e", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3914e94e-1c", "ovs_interfaceid": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1137.497365] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.515336] env[65107]: DEBUG nova.compute.manager [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1137.515607] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.516876] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92642a76-f056-47dd-8480-b7b6a1744a60 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.527697] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.528134] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eccf6d1-b431-4287-bca2-5f62b03454b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.538736] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1137.538736] env[65107]: value = "task-5103627" [ 1137.538736] env[65107]: _type = "Task" [ 1137.538736] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.557351] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.597581] env[65107]: INFO nova.compute.manager [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Took 19.88 seconds to build instance. [ 1137.605948] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103625, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.610310] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103626, 'name': Rename_Task, 'duration_secs': 0.183332} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.611103] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1137.611385] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61f421b0-c5cc-45e6-9a54-1cab074d1280 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.619350] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1137.619350] env[65107]: value = "task-5103628" [ 1137.619350] env[65107]: _type = "Task" [ 1137.619350] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.633026] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.668211] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523688c2-536d-1464-4c7d-a514c213d54f, 'name': SearchDatastore_Task, 'duration_secs': 0.010952} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.668424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.668752] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3/3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1137.668752] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae930842-5f33-4127-9198-99c918eb2121 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.678345] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1137.678345] env[65107]: value = "task-5103629" [ 1137.678345] env[65107]: _type = "Task" [ 1137.678345] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.687403] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103629, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.700053] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1137.833754] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a190ac-2d70-4a59-8bdf-59cdaa3c1756 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.841771] env[65107]: DEBUG oslo_concurrency.lockutils [req-2f6d17e5-ed69-46c4-9c36-c0a7d129d8c5 req-608e5a4e-984b-415f-bacf-a488b8610634 service nova] Releasing lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.843093] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e27a86-6017-4189-aa4c-2ce809331390 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.879657] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef429ad-3bb8-44e1-b221-8fba9c1cb5e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.889640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e65c5-dab2-4eb4-a2d1-b6e760729f3a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.906801] env[65107]: DEBUG nova.compute.provider_tree [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.996417] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.058640] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103627, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.091168] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103625, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.111114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7209d329-3fec-4c19-b350-dca0fe4e2b42 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.402s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.136020] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103628, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.195972] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103629, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.239596] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.411448] env[65107]: DEBUG nova.scheduler.client.report [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1138.499633] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.558398] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103627, 'name': PowerOffVM_Task, 'duration_secs': 0.568363} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.558852] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.559080] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1138.559441] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a88af09-9ed9-42c7-ba06-e4796aae11e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.589982] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103625, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.633974] env[65107]: DEBUG oslo_vmware.api [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103628, 'name': PowerOnVM_Task, 'duration_secs': 0.66393} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.634371] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1138.634621] env[65107]: INFO nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Took 9.37 seconds to spawn the instance on the hypervisor. [ 1138.634860] env[65107]: DEBUG nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1138.635809] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122ea7b7-e40e-47a5-ac34-cb1ba40e140e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.663787] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1138.664192] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1138.664443] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore1] 3f544b3f-703a-49d5-b3e8-16672766f691 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.664896] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74508e23-1fe0-42c8-9459-1171e462c278 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.674035] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1138.674035] env[65107]: value = "task-5103631" [ 1138.674035] env[65107]: _type = "Task" [ 1138.674035] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.683702] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103631, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.693232] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103629, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59232} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.693685] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3/3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1138.693976] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1138.694265] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfbda4e5-fd82-4c3f-b155-abed769add25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.704992] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1138.704992] env[65107]: value = "task-5103632" [ 1138.704992] env[65107]: _type = "Task" [ 1138.704992] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.717098] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.918093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.867s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.918655] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1138.922013] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.312s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.922344] env[65107]: DEBUG nova.objects.instance [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lazy-loading 'resources' on Instance uuid ee50d08a-57fc-4c05-96fe-a11fe2708165 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.996539] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.017019] env[65107]: DEBUG nova.compute.manager [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1139.017241] env[65107]: DEBUG nova.compute.manager [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing instance network info cache due to event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1139.017454] env[65107]: DEBUG oslo_concurrency.lockutils [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.017650] env[65107]: DEBUG oslo_concurrency.lockutils [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1139.017749] env[65107]: DEBUG nova.network.neutron [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1139.022560] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1139.049714] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.049973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.089296] env[65107]: DEBUG oslo_vmware.api [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103625, 'name': PowerOnVM_Task, 'duration_secs': 1.608546} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.089554] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1139.089761] env[65107]: INFO nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Took 14.77 seconds to spawn the instance on the hypervisor. [ 1139.089936] env[65107]: DEBUG nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1139.090743] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06aa1fcf-a220-496f-8750-b5dfef66e346 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.157693] env[65107]: INFO nova.compute.manager [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Took 20.67 seconds to build instance. [ 1139.185325] env[65107]: DEBUG oslo_vmware.api [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103631, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177073} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.185584] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1139.185783] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1139.185981] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1139.186162] env[65107]: INFO nova.compute.manager [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Took 1.67 seconds to destroy the instance on the hypervisor. [ 1139.186409] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1139.186600] env[65107]: DEBUG nova.compute.manager [-] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1139.186697] env[65107]: DEBUG nova.network.neutron [-] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1139.186937] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.187484] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.187750] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.216164] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.216376} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.216443] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1139.217319] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d70f16a-b31d-4953-8d5b-34c8269b72f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.245500] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3/3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.246837] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.249049] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75e6d79d-fe9f-4d30-9c1b-1edacdc5494b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.270523] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1139.270523] env[65107]: value = "task-5103633" [ 1139.270523] env[65107]: _type = "Task" [ 1139.270523] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.280201] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103633, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.430722] env[65107]: DEBUG nova.compute.utils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1139.432951] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1139.433244] env[65107]: DEBUG nova.network.neutron [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1139.433943] env[65107]: WARNING neutronclient.v2_0.client [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.434108] env[65107]: WARNING neutronclient.v2_0.client [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.434971] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.435494] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.506256] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.521784] env[65107]: WARNING neutronclient.v2_0.client [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.522537] env[65107]: WARNING openstack [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.522995] env[65107]: WARNING openstack [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.533041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.552957] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1139.609688] env[65107]: DEBUG nova.policy [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ded1c3ef07d4e78886dc924e9099363', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '782397ddd6cb40298dcf7f0da7353d67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1139.622162] env[65107]: INFO nova.compute.manager [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Took 22.21 seconds to build instance. [ 1139.659427] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b759a21-a8b1-4264-b5f7-af7f502e2562 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.175s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1139.667156] env[65107]: DEBUG nova.compute.manager [req-9d8d4232-f421-4d9e-bd62-0fa298bad33f req-a244b1fb-ea61-4639-bd8e-db141561f956 service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Received event network-vif-deleted-38203ad5-ce82-479b-8bb4-774b4fb04634 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1139.667329] env[65107]: INFO nova.compute.manager [req-9d8d4232-f421-4d9e-bd62-0fa298bad33f req-a244b1fb-ea61-4639-bd8e-db141561f956 service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Neutron deleted interface 38203ad5-ce82-479b-8bb4-774b4fb04634; detaching it from the instance and deleting it from the info cache [ 1139.668044] env[65107]: DEBUG nova.network.neutron [req-9d8d4232-f421-4d9e-bd62-0fa298bad33f req-a244b1fb-ea61-4639-bd8e-db141561f956 service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1139.783835] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.899494] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c56ca26-bc28-4055-abd0-a7bcfe0a0a83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.909237] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a016a501-7188-4da5-8744-24bd2f1f4357 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.948118] env[65107]: WARNING openstack [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.948526] env[65107]: WARNING openstack [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.955829] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cce70d7-1c7e-47c2-8ab7-f7bf9c22654b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.959278] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1139.968666] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b85a6e-9cdd-4869-9081-a01d262491b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.986458] env[65107]: DEBUG nova.compute.provider_tree [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.999028] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.007937] env[65107]: DEBUG nova.network.neutron [-] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1140.050170] env[65107]: DEBUG nova.network.neutron [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Successfully created port: e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1140.078525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.085451] env[65107]: WARNING neutronclient.v2_0.client [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.085451] env[65107]: WARNING openstack [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.085451] env[65107]: WARNING openstack [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.125891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19646250-1c92-44e2-8bd8-40af895be976 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.729s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.171734] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90738210-ea89-4b03-b1a7-7ba17c8c3900 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.184387] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8e01c0-66ac-4bf4-950b-fccd4bf0c2cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.223379] env[65107]: DEBUG nova.compute.manager [req-9d8d4232-f421-4d9e-bd62-0fa298bad33f req-a244b1fb-ea61-4639-bd8e-db141561f956 service nova] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Detach interface failed, port_id=38203ad5-ce82-479b-8bb4-774b4fb04634, reason: Instance 3f544b3f-703a-49d5-b3e8-16672766f691 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1140.233437] env[65107]: DEBUG nova.network.neutron [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updated VIF entry in instance network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1140.233796] env[65107]: DEBUG nova.network.neutron [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1140.285883] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103633, 'name': ReconfigVM_Task, 'duration_secs': 0.577437} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.286229] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3/3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1140.286944] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6757b063-f336-4cfa-adcc-37f833d3b3f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.295014] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1140.295014] env[65107]: value = "task-5103634" [ 1140.295014] env[65107]: _type = "Task" [ 1140.295014] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.304653] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103634, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.492166] env[65107]: DEBUG nova.scheduler.client.report [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.517436] env[65107]: INFO nova.compute.manager [-] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Took 1.33 seconds to deallocate network for instance. [ 1140.518183] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.736688] env[65107]: DEBUG oslo_concurrency.lockutils [req-6180fe58-2a5a-4a25-8aaf-134427f8d7dd req-9db6bfa6-f0fc-4950-835d-77e599b6f5ac service nova] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.806413] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103634, 'name': Rename_Task, 'duration_secs': 0.150572} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.806684] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1140.806934] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07a76ff7-6621-4539-825c-577e1b864271 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.814667] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1140.814667] env[65107]: value = "task-5103635" [ 1140.814667] env[65107]: _type = "Task" [ 1140.814667] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.824825] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.972767] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1140.988374] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "d52a955c-bc3c-41f2-b3ab-db9da613279e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.988923] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.007178] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.085s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.009440] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.011912] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1141.012187] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1141.012353] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1141.012535] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1141.012673] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1141.012852] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1141.013076] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1141.013236] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1141.013398] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1141.013556] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1141.013814] env[65107]: DEBUG nova.virt.hardware [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1141.015267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.126s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.015492] env[65107]: DEBUG nova.objects.instance [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lazy-loading 'resources' on Instance uuid 93cd2f0d-6bea-42c2-83b9-085dc94e27ab {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.017318] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00b44b8-1f29-4810-9f77-2b3090cdae10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.026368] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.031101] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5663e20-7234-4a8e-a7c0-4c41fff7f6d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.036482] env[65107]: INFO nova.scheduler.client.report [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Deleted allocations for instance ee50d08a-57fc-4c05-96fe-a11fe2708165 [ 1141.057070] env[65107]: DEBUG nova.compute.manager [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1141.057070] env[65107]: DEBUG nova.compute.manager [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing instance network info cache due to event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1141.057070] env[65107]: DEBUG oslo_concurrency.lockutils [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.057070] env[65107]: DEBUG oslo_concurrency.lockutils [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.057372] env[65107]: DEBUG nova.network.neutron [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1141.327673] env[65107]: DEBUG oslo_vmware.api [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103635, 'name': PowerOnVM_Task, 'duration_secs': 0.495479} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.327983] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1141.328230] env[65107]: INFO nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Took 7.75 seconds to spawn the instance on the hypervisor. [ 1141.328439] env[65107]: DEBUG nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1141.329438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a5e9e0-b280-448d-a8bf-108ec3d6c062 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.496524] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1141.503155] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.548605] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4a5ed9e9-f0c8-4835-b312-878be3ea6790 tempest-ServerRescueNegativeTestJSON-286642771 tempest-ServerRescueNegativeTestJSON-286642771-project-member] Lock "ee50d08a-57fc-4c05-96fe-a11fe2708165" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.640s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.562835] env[65107]: WARNING neutronclient.v2_0.client [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.563372] env[65107]: WARNING openstack [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.563735] env[65107]: WARNING openstack [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.647639] env[65107]: DEBUG nova.network.neutron [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Successfully updated port: e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1141.833716] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73bb406-293a-4bd1-a617-d8033123eeaa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.851604] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538e54a9-2143-454b-80b9-9ab999cde5a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.857101] env[65107]: INFO nova.compute.manager [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Took 17.86 seconds to build instance. [ 1141.891637] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59150c96-2eeb-4f8d-8c1c-652a13f4ad3f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.896508] env[65107]: WARNING openstack [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.897101] env[65107]: WARNING openstack [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.911924] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19931ac-be38-4fdf-8a2f-ae803ec83174 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.930080] env[65107]: DEBUG nova.compute.provider_tree [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.965984] env[65107]: WARNING neutronclient.v2_0.client [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.966795] env[65107]: WARNING openstack [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.967160] env[65107]: WARNING openstack [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.001022] env[65107]: DEBUG oslo_vmware.api [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103622, 'name': ReconfigVM_Task, 'duration_secs': 6.190968} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.003665] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.003665] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Reconfigured VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1142.003665] env[65107]: WARNING neutronclient.v2_0.client [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1142.003665] env[65107]: WARNING neutronclient.v2_0.client [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1142.003665] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.003665] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.029055] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.065105] env[65107]: WARNING neutronclient.v2_0.client [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1142.102232] env[65107]: DEBUG nova.network.neutron [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updated VIF entry in instance network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1142.103056] env[65107]: DEBUG nova.network.neutron [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.153409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.153555] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.153742] env[65107]: DEBUG nova.network.neutron [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1142.359842] env[65107]: DEBUG oslo_concurrency.lockutils [None req-08c23b09-5053-4741-aab1-0adbc65900ac tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.373s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.432904] env[65107]: DEBUG nova.scheduler.client.report [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.605918] env[65107]: DEBUG oslo_concurrency.lockutils [req-5c126e45-eec6-419c-81a2-47064c664551 req-35671353-295e-4214-aa77-074461ab6a68 service nova] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.657686] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.658319] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.694443] env[65107]: DEBUG nova.network.neutron [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1142.713758] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.714242] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.776783] env[65107]: WARNING neutronclient.v2_0.client [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1142.777274] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.777749] env[65107]: WARNING openstack [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.876752] env[65107]: DEBUG nova.network.neutron [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.939267] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.945209] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.704s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.947687] env[65107]: INFO nova.compute.claims [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.982775] env[65107]: INFO nova.scheduler.client.report [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Deleted allocations for instance 93cd2f0d-6bea-42c2-83b9-085dc94e27ab [ 1143.181037] env[65107]: DEBUG nova.compute.manager [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-vif-plugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1143.181037] env[65107]: DEBUG oslo_concurrency.lockutils [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.181037] env[65107]: DEBUG oslo_concurrency.lockutils [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1143.181878] env[65107]: DEBUG oslo_concurrency.lockutils [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.181878] env[65107]: DEBUG nova.compute.manager [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] No waiting events found dispatching network-vif-plugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1143.182039] env[65107]: WARNING nova.compute.manager [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received unexpected event network-vif-plugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede for instance with vm_state building and task_state spawning. [ 1143.182487] env[65107]: DEBUG nova.compute.manager [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-changed-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1143.182752] env[65107]: DEBUG nova.compute.manager [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Refreshing instance network info cache due to event network-changed-e48b19c2-6253-4a9c-a6fb-1daae3460ede. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1143.182950] env[65107]: DEBUG oslo_concurrency.lockutils [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.380502] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.381044] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance network_info: |[{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1143.381525] env[65107]: DEBUG oslo_concurrency.lockutils [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.381792] env[65107]: DEBUG nova.network.neutron [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Refreshing network info cache for port e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1143.383664] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:b0:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '17c839f5-4de0-449c-9a24-4e0e2fca37ca', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e48b19c2-6253-4a9c-a6fb-1daae3460ede', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1143.398402] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating folder: Project (782397ddd6cb40298dcf7f0da7353d67). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1143.399690] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fbce8152-7e4a-4aac-847d-0b3cf968b995 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.415612] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Created folder: Project (782397ddd6cb40298dcf7f0da7353d67) in parent group-v992574. [ 1143.415612] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating folder: Instances. Parent ref: group-v992876. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1143.415612] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-099367c0-ecae-4234-bbe4-3df9d828f070 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.419094] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.419411] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.419704] env[65107]: DEBUG nova.network.neutron [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1143.434604] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Created folder: Instances in parent group-v992876. [ 1143.435036] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1143.436491] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1143.436893] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7576ce1-c587-4c64-8e47-aa8f17340fed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.474097] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1143.474097] env[65107]: value = "task-5103638" [ 1143.474097] env[65107]: _type = "Task" [ 1143.474097] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.488334] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103638, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.493324] env[65107]: DEBUG oslo_concurrency.lockutils [None req-185b1a22-d0e4-4dcf-aac5-a4f7e52998df tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "93cd2f0d-6bea-42c2-83b9-085dc94e27ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.477s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.691836] env[65107]: DEBUG nova.compute.manager [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Received event network-changed-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1143.691836] env[65107]: DEBUG nova.compute.manager [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Refreshing instance network info cache due to event network-changed-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1143.691836] env[65107]: DEBUG oslo_concurrency.lockutils [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Acquiring lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.691836] env[65107]: DEBUG oslo_concurrency.lockutils [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Acquired lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.692141] env[65107]: DEBUG nova.network.neutron [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Refreshing network info cache for port 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1143.900891] env[65107]: WARNING neutronclient.v2_0.client [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.902291] env[65107]: WARNING openstack [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.902291] env[65107]: WARNING openstack [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.923223] env[65107]: WARNING neutronclient.v2_0.client [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.923976] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.924333] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.986257] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103638, 'name': CreateVM_Task, 'duration_secs': 0.409198} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.986455] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1143.987027] env[65107]: WARNING neutronclient.v2_0.client [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.987456] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.987629] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.987947] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1143.988252] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0ed6d39-b401-46e8-a756-8e4ff491e3bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.998467] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1143.998467] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5276ec78-9b5a-024b-174e-6fb0e4013c39" [ 1143.998467] env[65107]: _type = "Task" [ 1143.998467] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.010043] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5276ec78-9b5a-024b-174e-6fb0e4013c39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.197517] env[65107]: WARNING neutronclient.v2_0.client [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.199857] env[65107]: WARNING openstack [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.200430] env[65107]: WARNING openstack [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.243860] env[65107]: WARNING openstack [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.244282] env[65107]: WARNING openstack [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.319278] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.319744] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.435021] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613d4245-f71f-4abd-991b-07b98eda91f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.442937] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d8c6ca-cfc8-4e7a-863e-2efec84f73d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.479887] env[65107]: WARNING neutronclient.v2_0.client [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.480831] env[65107]: WARNING openstack [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.481374] env[65107]: WARNING openstack [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.489775] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc603c40-5a31-4917-85ef-9f60f4ebca69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.494084] env[65107]: WARNING neutronclient.v2_0.client [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.494501] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.494913] env[65107]: WARNING openstack [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.511337] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8465c951-d777-491e-a919-6b52bf2a7c0b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.520036] env[65107]: WARNING openstack [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.520430] env[65107]: WARNING openstack [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.527222] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5276ec78-9b5a-024b-174e-6fb0e4013c39, 'name': SearchDatastore_Task, 'duration_secs': 0.013269} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.528227] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.528467] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1144.528706] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.528881] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.529095] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1144.529385] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-189b294c-ee60-47b2-a135-214ffdfaf262 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.540016] env[65107]: DEBUG nova.compute.provider_tree [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.553438] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1144.553624] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1144.554448] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8b009f1-4066-4cb0-be9e-cb433bb5cb33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.561473] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1144.561473] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528db16e-d275-ea4e-8259-4efbc061e610" [ 1144.561473] env[65107]: _type = "Task" [ 1144.561473] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.576518] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528db16e-d275-ea4e-8259-4efbc061e610, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.615724] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.616235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.616383] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.616601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.616817] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.619693] env[65107]: INFO nova.compute.manager [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Terminating instance [ 1144.745220] env[65107]: DEBUG nova.network.neutron [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updated VIF entry in instance network info cache for port e48b19c2-6253-4a9c-a6fb-1daae3460ede. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1144.745365] env[65107]: DEBUG nova.network.neutron [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1144.779521] env[65107]: INFO nova.network.neutron [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Port 12711066-1ddb-4f73-aaf9-f2eb3d8c4515 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1144.779892] env[65107]: DEBUG nova.network.neutron [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1144.809467] env[65107]: WARNING neutronclient.v2_0.client [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.810212] env[65107]: WARNING openstack [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.810591] env[65107]: WARNING openstack [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.864237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-0599df85-ad33-4bd5-b234-611db1ef73f0-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.864349] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-0599df85-ad33-4bd5-b234-611db1ef73f0-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.865987] env[65107]: DEBUG nova.objects.instance [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'flavor' on Instance uuid 0599df85-ad33-4bd5-b234-611db1ef73f0 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.950022] env[65107]: DEBUG nova.network.neutron [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updated VIF entry in instance network info cache for port 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1144.950022] env[65107]: DEBUG nova.network.neutron [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updating instance_info_cache with network_info: [{"id": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "address": "fa:16:3e:f6:22:6e", "network": {"id": "859b6430-5f76-4a7b-bbfd-382116d2dc45", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-834115434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "544c649453844164bac53c98b342543f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3914e94e-1c", "ovs_interfaceid": "3914e94e-1c0c-410a-adcb-0ac7cb7ff3be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1145.044083] env[65107]: DEBUG nova.scheduler.client.report [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1145.073625] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528db16e-d275-ea4e-8259-4efbc061e610, 'name': SearchDatastore_Task, 'duration_secs': 0.027817} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.075324] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55f9e777-4a5a-4fdb-b862-0746df09e653 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.089019] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1145.089019] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a38a73-8470-dd8e-260e-ee31a1d043eb" [ 1145.089019] env[65107]: _type = "Task" [ 1145.089019] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.097912] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a38a73-8470-dd8e-260e-ee31a1d043eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.124480] env[65107]: DEBUG nova.compute.manager [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1145.127035] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1145.127035] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2528009-1f5e-4491-a3db-90ee240e36e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.138307] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1145.138886] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a3d322c-fbde-41f9-87e2-305c2ff9988b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.148049] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1145.148049] env[65107]: value = "task-5103639" [ 1145.148049] env[65107]: _type = "Task" [ 1145.148049] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.162183] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.224552] env[65107]: DEBUG nova.compute.manager [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1145.224552] env[65107]: DEBUG nova.compute.manager [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing instance network info cache due to event network-changed-47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1145.224552] env[65107]: DEBUG oslo_concurrency.lockutils [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Acquiring lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.248367] env[65107]: DEBUG oslo_concurrency.lockutils [req-b97dc969-6473-435a-958d-cb67b881a0cb req-5b240b41-94f0-49cd-9232-c0139b52a753 service nova] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.283555] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.286347] env[65107]: DEBUG oslo_concurrency.lockutils [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Acquired lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.286506] env[65107]: DEBUG nova.network.neutron [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Refreshing network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1145.370629] env[65107]: WARNING neutronclient.v2_0.client [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.371351] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.371736] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.453343] env[65107]: DEBUG oslo_concurrency.lockutils [req-a4c86498-5891-45bf-9a32-73d57aa55301 req-85ff99a5-bab0-4a1e-8495-cf48ce0e73fa service nova] Releasing lock "refresh_cache-3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.503780] env[65107]: DEBUG nova.objects.instance [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'pci_requests' on Instance uuid 0599df85-ad33-4bd5-b234-611db1ef73f0 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.549533] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.550089] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1145.554241] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.022s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1145.554698] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.554698] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1145.554929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.477s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1145.556366] env[65107]: INFO nova.compute.claims [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1145.559767] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19b2d2c-6400-41f2-8496-6599e0a71695 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.569310] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f3969f-dbc0-4b12-b0ec-ad5bffbf2c4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.586122] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b944c050-42cb-45fa-a1b0-f1a940c352b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.601366] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a38a73-8470-dd8e-260e-ee31a1d043eb, 'name': SearchDatastore_Task, 'duration_secs': 0.056086} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.601727] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.602020] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1145.603072] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adb8d41-927e-43dd-9580-1a0830def184 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.607203] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d34432d-8e99-463d-bb62-fd42ebd4fa9c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.641291] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178546MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1145.641521] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1145.643623] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1145.643623] env[65107]: value = "task-5103640" [ 1145.643623] env[65107]: _type = "Task" [ 1145.643623] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.656575] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.663201] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103639, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.790659] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0933e85d-cc2b-40ed-a718-5540e39a4a27 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.806s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.790659] env[65107]: WARNING neutronclient.v2_0.client [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.791333] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.791704] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.952240] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.952657] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.006907] env[65107]: DEBUG nova.objects.base [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Object Instance<0599df85-ad33-4bd5-b234-611db1ef73f0> lazy-loaded attributes: flavor,pci_requests {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1146.007676] env[65107]: DEBUG nova.network.neutron [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1146.007676] env[65107]: WARNING neutronclient.v2_0.client [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.007941] env[65107]: WARNING neutronclient.v2_0.client [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.008386] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.008762] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.061406] env[65107]: DEBUG nova.compute.utils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1146.062300] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1146.062463] env[65107]: DEBUG nova.network.neutron [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1146.062860] env[65107]: WARNING neutronclient.v2_0.client [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.063478] env[65107]: WARNING neutronclient.v2_0.client [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.064083] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.064436] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.086956] env[65107]: WARNING neutronclient.v2_0.client [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.087706] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.087992] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.158116] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103640, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.164433] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103639, 'name': PowerOffVM_Task, 'duration_secs': 0.576751} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.164610] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1146.164817] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1146.165153] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07af57e5-1bcd-49cd-a4da-16d5ffdeee98 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.184024] env[65107]: DEBUG nova.policy [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '365eed31877241e58fc860ed5620d5f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c805fe68d0fd4147b9b4b7fdc9d6a6d0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1146.261317] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1146.261961] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1146.261961] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Deleting the datastore file [datastore2] 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1146.262219] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21497f34-2ebc-4b15-8d21-95ed5596dae9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.273199] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for the task: (returnval){ [ 1146.273199] env[65107]: value = "task-5103642" [ 1146.273199] env[65107]: _type = "Task" [ 1146.273199] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.285695] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103642, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.296653] env[65107]: DEBUG nova.policy [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d321b88edf4587968ad36cbd9a041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '152868ee0723449baac10f6a8ceeb738', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1146.576380] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1146.587268] env[65107]: DEBUG nova.network.neutron [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updated VIF entry in instance network info cache for port 47322e57-eaa6-4bb8-98ec-3098749bd52c. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1146.587714] env[65107]: DEBUG nova.network.neutron [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [{"id": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "address": "fa:16:3e:84:11:bd", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47322e57-ea", "ovs_interfaceid": "47322e57-eaa6-4bb8-98ec-3098749bd52c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1146.663945] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.929694} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.668923] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1146.668923] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1146.668923] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00667041-a480-4fde-b719-369562d53035 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.678165] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1146.678165] env[65107]: value = "task-5103643" [ 1146.678165] env[65107]: _type = "Task" [ 1146.678165] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.689590] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.733155] env[65107]: DEBUG nova.network.neutron [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Successfully created port: 3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1146.788931] env[65107]: DEBUG oslo_vmware.api [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Task: {'id': task-5103642, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49622} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.789243] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1146.789431] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1146.789612] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1146.789784] env[65107]: INFO nova.compute.manager [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Took 1.67 seconds to destroy the instance on the hypervisor. [ 1146.790092] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1146.790299] env[65107]: DEBUG nova.compute.manager [-] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1146.790397] env[65107]: DEBUG nova.network.neutron [-] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1146.790629] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.791209] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.791432] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.918330] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba7fc50-309a-4252-994e-337a927cbd90 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.926909] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a08683-e0e6-4fb6-9ded-7d304db3268a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.961875] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def8c21a-4b22-413b-a2e9-2d51927b0235 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.970791] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db92355e-41d5-41e0-a552-eecc3757711c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.986501] env[65107]: DEBUG nova.compute.provider_tree [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.094694] env[65107]: DEBUG oslo_concurrency.lockutils [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Releasing lock "refresh_cache-709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.094997] env[65107]: DEBUG nova.compute.manager [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1147.095200] env[65107]: DEBUG nova.compute.manager [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing instance network info cache due to event network-changed-d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1147.095415] env[65107]: DEBUG oslo_concurrency.lockutils [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.095554] env[65107]: DEBUG oslo_concurrency.lockutils [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1147.095758] env[65107]: DEBUG nova.network.neutron [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1147.189298] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074593} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.189631] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1147.190506] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec21618-2a8c-492a-b618-ad624ac92783 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.215486] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.215875] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fabe39f-901f-4187-a5f4-6b97a327aea6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.238843] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1147.238843] env[65107]: value = "task-5103644" [ 1147.238843] env[65107]: _type = "Task" [ 1147.238843] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.248495] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103644, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.490259] env[65107]: DEBUG nova.scheduler.client.report [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1147.588081] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1147.600770] env[65107]: WARNING neutronclient.v2_0.client [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.600770] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.601211] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.617538] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1147.618243] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1147.618243] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1147.618243] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1147.618402] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1147.618516] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1147.618721] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.618930] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1147.619066] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1147.619606] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1147.619606] env[65107]: DEBUG nova.virt.hardware [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1147.620275] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee86043-00e8-4611-978c-7873bca13d5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.630053] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efad060f-e259-4693-a35a-6403a463a0b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.749657] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103644, 'name': ReconfigVM_Task, 'duration_secs': 0.319014} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.749923] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Reconfigured VM instance instance-00000071 to attach disk [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1147.750695] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1d3161e-26cf-43cc-813d-8b9bd3fdd02f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.758752] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1147.758752] env[65107]: value = "task-5103645" [ 1147.758752] env[65107]: _type = "Task" [ 1147.758752] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.768213] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103645, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.996080] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.996744] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1148.001931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.973s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.001931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.002237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.973s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.006259] env[65107]: INFO nova.compute.claims [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1148.043315] env[65107]: INFO nova.scheduler.client.report [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocations for instance 3f544b3f-703a-49d5-b3e8-16672766f691 [ 1148.279532] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103645, 'name': Rename_Task, 'duration_secs': 0.183205} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.279532] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1148.279532] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea8c9475-4e92-4e7e-9324-db760f9dff4a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.290088] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1148.290088] env[65107]: value = "task-5103646" [ 1148.290088] env[65107]: _type = "Task" [ 1148.290088] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.299349] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.329086] env[65107]: DEBUG nova.network.neutron [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Successfully updated port: 3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1148.402609] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.511994] env[65107]: DEBUG nova.compute.utils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1148.517675] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1148.522573] env[65107]: DEBUG nova.network.neutron [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1148.522573] env[65107]: WARNING neutronclient.v2_0.client [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.522573] env[65107]: WARNING neutronclient.v2_0.client [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.522902] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.524568] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.554961] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b2a82482-6a87-4f33-a630-c5bbb697fff4 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "3f544b3f-703a-49d5-b3e8-16672766f691" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.583s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.625038] env[65107]: DEBUG nova.network.neutron [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Successfully updated port: 12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1148.724331] env[65107]: DEBUG nova.policy [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b73837038b2c4e678c0a7112f6a61f38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd28d5a8ab6f747628edf3aee83355feb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1148.774067] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.774067] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.799600] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103646, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.834266] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "refresh_cache-0161c1ba-8aff-4f70-a407-caa991e43f15" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.834516] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "refresh_cache-0161c1ba-8aff-4f70-a407-caa991e43f15" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1148.834999] env[65107]: DEBUG nova.network.neutron [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1149.018785] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1149.073550] env[65107]: DEBUG nova.network.neutron [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Successfully created port: 079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1149.085324] env[65107]: WARNING neutronclient.v2_0.client [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1149.086126] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1149.086548] env[65107]: WARNING openstack [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1149.128342] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.185021] env[65107]: DEBUG nova.network.neutron [-] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1149.304757] env[65107]: DEBUG oslo_vmware.api [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103646, 'name': PowerOnVM_Task, 'duration_secs': 0.688605} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.305085] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1149.305440] env[65107]: INFO nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Took 8.33 seconds to spawn the instance on the hypervisor. [ 1149.305726] env[65107]: DEBUG nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1149.306759] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd858edc-e141-446e-af15-44c1901b5b8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.340585] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1149.340995] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1149.358523] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c088030-945b-4e12-a983-3284295aa82f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.368094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1893b8-05b7-4ca1-bd2a-63e05f8b6e50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.405589] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49932431-b496-4ed3-88c4-55fda660d120 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.415343] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18360f4-e384-4e8f-af80-1e29d6a0f2d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.431598] env[65107]: DEBUG nova.compute.provider_tree [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.673097] env[65107]: DEBUG nova.network.neutron [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1149.687461] env[65107]: INFO nova.compute.manager [-] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Took 2.90 seconds to deallocate network for instance. [ 1149.713457] env[65107]: DEBUG nova.network.neutron [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updated VIF entry in instance network info cache for port d04b7a1c-d3b0-4211-b44b-2e39791f7497. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1149.713861] env[65107]: DEBUG nova.network.neutron [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1149.758108] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1149.758500] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1149.834215] env[65107]: WARNING neutronclient.v2_0.client [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1149.834954] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1149.835298] env[65107]: WARNING openstack [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1149.843014] env[65107]: INFO nova.compute.manager [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Took 16.89 seconds to build instance. [ 1149.934989] env[65107]: DEBUG nova.scheduler.client.report [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1149.976371] env[65107]: DEBUG nova.compute.manager [req-f1e5b038-1ec7-40c2-af62-93fa9462ec00 req-5526b9ad-0c6f-4c42-95cc-9ccbb99be38d service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-vif-plugged-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1149.976597] env[65107]: DEBUG oslo_concurrency.lockutils [req-f1e5b038-1ec7-40c2-af62-93fa9462ec00 req-5526b9ad-0c6f-4c42-95cc-9ccbb99be38d service nova] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.976814] env[65107]: DEBUG oslo_concurrency.lockutils [req-f1e5b038-1ec7-40c2-af62-93fa9462ec00 req-5526b9ad-0c6f-4c42-95cc-9ccbb99be38d service nova] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.976986] env[65107]: DEBUG oslo_concurrency.lockutils [req-f1e5b038-1ec7-40c2-af62-93fa9462ec00 req-5526b9ad-0c6f-4c42-95cc-9ccbb99be38d service nova] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.977171] env[65107]: DEBUG nova.compute.manager [req-f1e5b038-1ec7-40c2-af62-93fa9462ec00 req-5526b9ad-0c6f-4c42-95cc-9ccbb99be38d service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] No waiting events found dispatching network-vif-plugged-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1149.977335] env[65107]: WARNING nova.compute.manager [req-f1e5b038-1ec7-40c2-af62-93fa9462ec00 req-5526b9ad-0c6f-4c42-95cc-9ccbb99be38d service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received unexpected event network-vif-plugged-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 for instance with vm_state active and task_state None. [ 1149.985952] env[65107]: DEBUG nova.network.neutron [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Updating instance_info_cache with network_info: [{"id": "3e149eec-773a-449a-8c36-4dd31d98e502", "address": "fa:16:3e:7f:5e:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e149eec-77", "ovs_interfaceid": "3e149eec-773a-449a-8c36-4dd31d98e502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1149.994817] env[65107]: DEBUG nova.compute.manager [req-1994677c-2872-41da-af07-c88ef31fc1af req-3cc7b00d-e696-4f85-9434-9d3b50405939 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Received event network-vif-plugged-3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1149.996581] env[65107]: DEBUG oslo_concurrency.lockutils [req-1994677c-2872-41da-af07-c88ef31fc1af req-3cc7b00d-e696-4f85-9434-9d3b50405939 service nova] Acquiring lock "0161c1ba-8aff-4f70-a407-caa991e43f15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.996581] env[65107]: DEBUG oslo_concurrency.lockutils [req-1994677c-2872-41da-af07-c88ef31fc1af req-3cc7b00d-e696-4f85-9434-9d3b50405939 service nova] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.996581] env[65107]: DEBUG oslo_concurrency.lockutils [req-1994677c-2872-41da-af07-c88ef31fc1af req-3cc7b00d-e696-4f85-9434-9d3b50405939 service nova] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.996581] env[65107]: DEBUG nova.compute.manager [req-1994677c-2872-41da-af07-c88ef31fc1af req-3cc7b00d-e696-4f85-9434-9d3b50405939 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] No waiting events found dispatching network-vif-plugged-3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1149.996581] env[65107]: WARNING nova.compute.manager [req-1994677c-2872-41da-af07-c88ef31fc1af req-3cc7b00d-e696-4f85-9434-9d3b50405939 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Received unexpected event network-vif-plugged-3e149eec-773a-449a-8c36-4dd31d98e502 for instance with vm_state building and task_state spawning. [ 1150.030786] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1150.058530] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1150.058804] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1150.058961] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1150.059162] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1150.059307] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1150.059493] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1150.059734] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1150.059896] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1150.060143] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1150.060361] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1150.060544] env[65107]: DEBUG nova.virt.hardware [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1150.061522] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b31592-53ad-4f02-8ce9-e868344a0181 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.071485] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456c8abe-3645-449a-87b7-a9e8baa5bd78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.194074] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1150.217195] env[65107]: DEBUG oslo_concurrency.lockutils [req-8abe9369-0817-4f32-932a-23dc4e10139c req-845e308a-53b3-4d5f-a3fc-3505bed18ed5 service nova] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1150.217677] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.217876] env[65107]: DEBUG nova.network.neutron [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1150.345552] env[65107]: DEBUG oslo_concurrency.lockutils [None req-31d4e266-3484-470a-a18d-916db35fb171 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.409s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1150.442322] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1150.443097] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1150.445886] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.804s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1150.489670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "refresh_cache-0161c1ba-8aff-4f70-a407-caa991e43f15" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1150.490091] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Instance network_info: |[{"id": "3e149eec-773a-449a-8c36-4dd31d98e502", "address": "fa:16:3e:7f:5e:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e149eec-77", "ovs_interfaceid": "3e149eec-773a-449a-8c36-4dd31d98e502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1150.490874] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:5e:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e149eec-773a-449a-8c36-4dd31d98e502', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1150.499068] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1150.500084] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1150.501118] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b1bd7e7-4c81-4186-8dc6-a1b2e0616772 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.525518] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1150.525518] env[65107]: value = "task-5103647" [ 1150.525518] env[65107]: _type = "Task" [ 1150.525518] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.536243] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103647, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.721034] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.721484] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.741162] env[65107]: DEBUG nova.network.neutron [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Successfully updated port: 079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1150.771304] env[65107]: WARNING nova.network.neutron [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] 40a76aa5-ac46-4067-a98d-cd60410bf479 already exists in list: networks containing: ['40a76aa5-ac46-4067-a98d-cd60410bf479']. ignoring it [ 1150.813663] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.814074] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.898616] env[65107]: WARNING neutronclient.v2_0.client [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.899329] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.899671] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.948994] env[65107]: DEBUG nova.compute.utils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1150.951443] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1150.951768] env[65107]: DEBUG nova.network.neutron [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1150.952111] env[65107]: WARNING neutronclient.v2_0.client [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.952502] env[65107]: WARNING neutronclient.v2_0.client [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.953034] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.953370] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.016777] env[65107]: DEBUG nova.policy [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1151.022277] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.022658] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.041141] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103647, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.096626] env[65107]: WARNING neutronclient.v2_0.client [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.097684] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.098158] env[65107]: WARNING openstack [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.198840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.199098] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.202355] env[65107]: DEBUG nova.network.neutron [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "address": "fa:16:3e:ec:8b:30", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12711066-1d", "ovs_interfaceid": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1151.244148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "refresh_cache-4f93a52c-9c40-47e5-a34f-b19aea34d7a9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.244416] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "refresh_cache-4f93a52c-9c40-47e5-a34f-b19aea34d7a9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.244970] env[65107]: DEBUG nova.network.neutron [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1151.358360] env[65107]: DEBUG nova.network.neutron [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Successfully created port: b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1151.452448] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1151.505425] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.505425] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3df65e64-0c45-4707-960e-8f1767e2d011 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.505425] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 2c45d232-a948-4ad7-80a0-589c2dc91ec1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.505621] env[65107]: WARNING nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1151.505705] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.505773] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 0599df85-ad33-4bd5-b234-611db1ef73f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.505966] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d7b4bc1e-c94e-4654-9345-2a8aa945a896 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506057] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bef91b92-add1-4df9-bc24-dab15ce04338 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506104] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 1f3292b7-f6d7-43ca-9caf-8c55794e8311 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 93a1b356-b6d5-435b-b30f-ea5ade031b7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 475851f6-24a2-4f0e-9567-d2a23947a5e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bcced66b-1464-4834-892b-3d0e652da31a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 0161c1ba-8aff-4f70-a407-caa991e43f15 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 4f93a52c-9c40-47e5-a34f-b19aea34d7a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.506895] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d52a955c-bc3c-41f2-b3ab-db9da613279e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1151.540348] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103647, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.554090] env[65107]: INFO nova.compute.manager [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Rebuilding instance [ 1151.611569] env[65107]: DEBUG nova.compute.manager [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1151.612509] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8536400d-0a26-4025-b96b-ce01f61fc854 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.704547] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1151.709430] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.709430] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.709430] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.710138] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a03f99-b028-4563-b383-777f480dc65f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.729410] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1151.729841] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1151.729841] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1151.729983] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1151.730149] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1151.730315] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1151.730505] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.730673] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1151.730828] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1151.730987] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1151.731265] env[65107]: DEBUG nova.virt.hardware [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1151.738714] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Reconfiguring VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1151.739133] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-175b91db-9d02-496f-b973-e62bb714166c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.752824] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.753242] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.769768] env[65107]: DEBUG oslo_vmware.api [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1151.769768] env[65107]: value = "task-5103648" [ 1151.769768] env[65107]: _type = "Task" [ 1151.769768] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.779347] env[65107]: DEBUG oslo_vmware.api [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103648, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.811614] env[65107]: DEBUG nova.network.neutron [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1151.841435] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.841781] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.921963] env[65107]: WARNING neutronclient.v2_0.client [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.922671] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.923042] env[65107]: WARNING openstack [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.010503] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 751ebb31-1a27-4630-9213-4312a6fa9298 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1152.011372] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1152.011372] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3392MB phys_disk=100GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '15', 'num_vm_active': '12', 'num_task_None': '13', 'num_os_type_None': '15', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '2', 'io_workload': '3', 'num_proj_7784762e36ae4c6fb30012ec801c3b88': '2', 'num_proj_645feeb0c8524267bff18c5f37f5f011': '3', 'num_proj_152868ee0723449baac10f6a8ceeb738': '2', 'num_proj_f379144b78764fe394039d87b043a946': '1', 'num_proj_dced4f9aef2f49cf990203b693533aa1': '1', 'num_proj_544c649453844164bac53c98b342543f': '1', 'num_proj_782397ddd6cb40298dcf7f0da7353d67': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_c805fe68d0fd4147b9b4b7fdc9d6a6d0': '1', 'num_proj_d28d5a8ab6f747628edf3aee83355feb': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1152.018806] env[65107]: DEBUG nova.compute.manager [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-changed-12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.018898] env[65107]: DEBUG nova.compute.manager [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing instance network info cache due to event network-changed-12711066-1ddb-4f73-aaf9-f2eb3d8c4515. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1152.019656] env[65107]: DEBUG oslo_concurrency.lockutils [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.019852] env[65107]: DEBUG oslo_concurrency.lockutils [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.020030] env[65107]: DEBUG nova.network.neutron [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Refreshing network info cache for port 12711066-1ddb-4f73-aaf9-f2eb3d8c4515 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1152.023507] env[65107]: DEBUG nova.network.neutron [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Updating instance_info_cache with network_info: [{"id": "079f6503-3319-4d8a-b94b-843464a762f0", "address": "fa:16:3e:7b:a0:24", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap079f6503-33", "ovs_interfaceid": "079f6503-3319-4d8a-b94b-843464a762f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1152.049772] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103647, 'name': CreateVM_Task, 'duration_secs': 1.418413} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.051671] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Received event network-changed-3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.051871] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Refreshing instance network info cache due to event network-changed-3e149eec-773a-449a-8c36-4dd31d98e502. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1152.052098] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Acquiring lock "refresh_cache-0161c1ba-8aff-4f70-a407-caa991e43f15" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.052240] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Acquired lock "refresh_cache-0161c1ba-8aff-4f70-a407-caa991e43f15" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.052396] env[65107]: DEBUG nova.network.neutron [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Refreshing network info cache for port 3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1152.053468] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1152.057522] env[65107]: WARNING neutronclient.v2_0.client [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.057885] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.058053] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.058429] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1152.059067] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9b82425-1eed-40c7-b2d2-590c0fece8b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.066397] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1152.066397] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5275a47e-a89d-b36e-f73a-560c3e891e65" [ 1152.066397] env[65107]: _type = "Task" [ 1152.066397] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.080406] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5275a47e-a89d-b36e-f73a-560c3e891e65, 'name': SearchDatastore_Task, 'duration_secs': 0.012112} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.080758] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.081029] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1152.081685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.081685] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.081685] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1152.082419] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1bdf3e4-1bcc-4e91-9280-0282949d1c8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.095511] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1152.095695] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1152.096638] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c87dae2-a9b4-4f05-85a2-d0b370c503b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.104076] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1152.104076] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529248e4-6670-f75b-351e-7fb0a6e3c9e5" [ 1152.104076] env[65107]: _type = "Task" [ 1152.104076] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.117529] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]529248e4-6670-f75b-351e-7fb0a6e3c9e5, 'name': SearchDatastore_Task} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.118421] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cc628f5-687f-4997-b32a-2e10fe87bd38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.128590] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1152.128590] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5228e8dd-4245-fd42-72fd-c39bc5903990" [ 1152.128590] env[65107]: _type = "Task" [ 1152.128590] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.137241] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "7e72891b-4c20-4889-80c9-8bcb15373a2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.137473] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.148949] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5228e8dd-4245-fd42-72fd-c39bc5903990, 'name': SearchDatastore_Task, 'duration_secs': 0.010749} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.149241] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.149500] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0161c1ba-8aff-4f70-a407-caa991e43f15/0161c1ba-8aff-4f70-a407-caa991e43f15.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1152.149769] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea253007-2a2c-43bd-8e02-af8a0d806c3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.161923] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1152.161923] env[65107]: value = "task-5103649" [ 1152.161923] env[65107]: _type = "Task" [ 1152.161923] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.172184] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103649, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.230752] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.283821] env[65107]: DEBUG oslo_vmware.api [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103648, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.312903] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf22625-d061-4958-9be2-08f997f8460a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.322827] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af2db13-5603-4326-8a12-c8b0251231fd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.357035] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc1311c-c22e-4ab2-9f81-ba91f4c7c59e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.365697] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c32e96-1b32-4925-b869-31d9315ae8b7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.383207] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.462064] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1152.494481] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1152.494861] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1152.495044] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1152.495304] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1152.495456] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1152.495695] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1152.495903] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1152.496093] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1152.496250] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1152.496411] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1152.496584] env[65107]: DEBUG nova.virt.hardware [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1152.497550] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65fbac1-a4d8-45a1-b4f3-7e4c8722364d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.508077] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bd6f97-bc71-4b3e-9399-0fba1b6ba690 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.526748] env[65107]: WARNING neutronclient.v2_0.client [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.527464] env[65107]: WARNING openstack [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.527868] env[65107]: WARNING openstack [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.535462] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "refresh_cache-4f93a52c-9c40-47e5-a34f-b19aea34d7a9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.535850] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Instance network_info: |[{"id": "079f6503-3319-4d8a-b94b-843464a762f0", "address": "fa:16:3e:7b:a0:24", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap079f6503-33", "ovs_interfaceid": "079f6503-3319-4d8a-b94b-843464a762f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1152.536654] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:a0:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '079f6503-3319-4d8a-b94b-843464a762f0', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1152.544675] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1152.545542] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1152.545542] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7c03322-c854-47f0-bfb1-d7251593bb7a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.560575] env[65107]: WARNING neutronclient.v2_0.client [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.561329] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.561702] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.580807] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1152.580807] env[65107]: value = "task-5103650" [ 1152.580807] env[65107]: _type = "Task" [ 1152.580807] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.594025] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103650, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.631794] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1152.632166] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3eebe895-180c-4099-9f29-4dac90c52be4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.642878] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1152.646977] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1152.646977] env[65107]: value = "task-5103651" [ 1152.646977] env[65107]: _type = "Task" [ 1152.646977] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.659755] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103651, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.677758] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103649, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.768212] env[65107]: WARNING openstack [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.768713] env[65107]: WARNING openstack [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.787207] env[65107]: DEBUG oslo_vmware.api [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103648, 'name': ReconfigVM_Task, 'duration_secs': 0.995116} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.788248] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.788606] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.797016] env[65107]: WARNING neutronclient.v2_0.client [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.797287] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.797660] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Reconfigured VM to attach interface {{(pid=65107) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1152.887483] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1152.911374] env[65107]: WARNING neutronclient.v2_0.client [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.912200] env[65107]: WARNING openstack [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.912629] env[65107]: WARNING openstack [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.969794] env[65107]: WARNING neutronclient.v2_0.client [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.970514] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.970872] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.979939] env[65107]: DEBUG nova.network.neutron [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Successfully updated port: b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1153.049471] env[65107]: DEBUG nova.network.neutron [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updated VIF entry in instance network info cache for port 12711066-1ddb-4f73-aaf9-f2eb3d8c4515. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1153.049991] env[65107]: DEBUG nova.network.neutron [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "address": "fa:16:3e:ec:8b:30", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12711066-1d", "ovs_interfaceid": "12711066-1ddb-4f73-aaf9-f2eb3d8c4515", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.088523] env[65107]: DEBUG nova.network.neutron [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Updated VIF entry in instance network info cache for port 3e149eec-773a-449a-8c36-4dd31d98e502. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1153.088884] env[65107]: DEBUG nova.network.neutron [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Updating instance_info_cache with network_info: [{"id": "3e149eec-773a-449a-8c36-4dd31d98e502", "address": "fa:16:3e:7f:5e:18", "network": {"id": "fac10a6e-add1-4dc7-b599-98befce60db0", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1003041336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c805fe68d0fd4147b9b4b7fdc9d6a6d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e149eec-77", "ovs_interfaceid": "3e149eec-773a-449a-8c36-4dd31d98e502", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.093790] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103650, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.168875] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103651, 'name': PowerOffVM_Task, 'duration_secs': 0.268937} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.173988] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1153.174093] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1153.175526] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.176773] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6dd761-0312-43c9-b71e-17141040bd7f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.190040] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103649, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528648} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.194026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 0161c1ba-8aff-4f70-a407-caa991e43f15/0161c1ba-8aff-4f70-a407-caa991e43f15.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1153.194259] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1153.194699] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1153.195276] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d13fd9aa-dba8-41cd-b64b-f41fc3189443 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.199546] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3f939a7-c7fc-40e4-8c22-320dfa2a3cdb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.210296] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1153.210296] env[65107]: value = "task-5103653" [ 1153.210296] env[65107]: _type = "Task" [ 1153.210296] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.221732] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.303316] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8a2c6877-c893-4a37-8ed9-2e4fa11be147 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-0599df85-ad33-4bd5-b234-611db1ef73f0-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.439s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.313229] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1153.313229] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1153.313229] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleting the datastore file [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.313687] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6bd7d6d-967f-4035-84ab-b75b2766e0be {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.321131] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1153.321131] env[65107]: value = "task-5103654" [ 1153.321131] env[65107]: _type = "Task" [ 1153.321131] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.332800] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.394281] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1153.394655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.949s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.395319] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.201s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.395515] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.397805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.167s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.399755] env[65107]: INFO nova.compute.claims [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1153.402674] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.403803] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Cleaning up deleted instances {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 1153.431540] env[65107]: INFO nova.scheduler.client.report [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Deleted allocations for instance 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3 [ 1153.484488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-d52a955c-bc3c-41f2-b3ab-db9da613279e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.484488] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-d52a955c-bc3c-41f2-b3ab-db9da613279e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.484488] env[65107]: DEBUG nova.network.neutron [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1153.555892] env[65107]: DEBUG oslo_concurrency.lockutils [req-b1bb1550-da69-494d-96e0-341a97dc6772 req-d74e7a68-d993-4aca-8899-75aa9a2d06a9 service nova] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.595702] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Releasing lock "refresh_cache-0161c1ba-8aff-4f70-a407-caa991e43f15" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.596234] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Received event network-vif-deleted-5343cf6f-59a3-4f9c-a534-449b9816168a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1153.596498] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Received event network-vif-plugged-079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1153.596754] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Acquiring lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.597040] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.597259] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.597438] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] No waiting events found dispatching network-vif-plugged-079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1153.597607] env[65107]: WARNING nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Received unexpected event network-vif-plugged-079f6503-3319-4d8a-b94b-843464a762f0 for instance with vm_state building and task_state spawning. [ 1153.597770] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Received event network-changed-079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1153.597924] env[65107]: DEBUG nova.compute.manager [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Refreshing instance network info cache due to event network-changed-079f6503-3319-4d8a-b94b-843464a762f0. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1153.598130] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Acquiring lock "refresh_cache-4f93a52c-9c40-47e5-a34f-b19aea34d7a9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.598367] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Acquired lock "refresh_cache-4f93a52c-9c40-47e5-a34f-b19aea34d7a9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.598665] env[65107]: DEBUG nova.network.neutron [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Refreshing network info cache for port 079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1153.600141] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103650, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.721451] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.178739} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.721813] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1153.722819] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62146eb1-26d6-4069-bc1f-c6a25227670d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.747932] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 0161c1ba-8aff-4f70-a407-caa991e43f15/0161c1ba-8aff-4f70-a407-caa991e43f15.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1153.748274] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5b5f416-cfb2-4c14-abe2-edb1af4f64ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.770370] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1153.770370] env[65107]: value = "task-5103655" [ 1153.770370] env[65107]: _type = "Task" [ 1153.770370] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.779847] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103655, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.831456] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.441988} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.831717] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.831911] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1153.832096] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1153.922773] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] There are 63 instances to clean {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 1153.922962] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: c2bbaa63-dc53-432d-bcaf-23e698f1f02a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1153.944220] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6f5604ad-694c-4cfa-9983-a91d33dd9911 tempest-ServerRescueTestJSON-675599107 tempest-ServerRescueTestJSON-675599107-project-member] Lock "23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.328s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.990628] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.990684] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.031069] env[65107]: DEBUG nova.network.neutron [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1154.060405] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.060946] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.099208] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103650, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.101300] env[65107]: WARNING neutronclient.v2_0.client [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.101957] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.102317] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.118321] env[65107]: DEBUG nova.compute.manager [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Received event network-vif-plugged-b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1154.118535] env[65107]: DEBUG oslo_concurrency.lockutils [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Acquiring lock "d52a955c-bc3c-41f2-b3ab-db9da613279e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.118736] env[65107]: DEBUG oslo_concurrency.lockutils [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.118925] env[65107]: DEBUG oslo_concurrency.lockutils [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.119103] env[65107]: DEBUG nova.compute.manager [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] No waiting events found dispatching network-vif-plugged-b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1154.119509] env[65107]: WARNING nova.compute.manager [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Received unexpected event network-vif-plugged-b633cd1f-f333-4c16-8e85-b43212472666 for instance with vm_state building and task_state spawning. [ 1154.119682] env[65107]: DEBUG nova.compute.manager [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Received event network-changed-b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1154.119835] env[65107]: DEBUG nova.compute.manager [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Refreshing instance network info cache due to event network-changed-b633cd1f-f333-4c16-8e85-b43212472666. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1154.120008] env[65107]: DEBUG oslo_concurrency.lockutils [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Acquiring lock "refresh_cache-d52a955c-bc3c-41f2-b3ab-db9da613279e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.188287] env[65107]: WARNING neutronclient.v2_0.client [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.189147] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.189601] env[65107]: WARNING openstack [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.251096] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b8c184-19d4-4ba5-8bd7-02f9d3409964 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.258483] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.258796] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.269091] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac5d7ff-8547-40fd-a5b1-60eb51002c50 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.312842] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103655, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.319098] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0340d15-71e2-4e88-a69f-5d38dc907613 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.329265] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47546b2a-62f3-41c3-8ef9-7a94bea7eed5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.352925] env[65107]: DEBUG nova.compute.provider_tree [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.367072] env[65107]: DEBUG nova.network.neutron [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Updating instance_info_cache with network_info: [{"id": "b633cd1f-f333-4c16-8e85-b43212472666", "address": "fa:16:3e:f7:a0:37", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb633cd1f-f3", "ovs_interfaceid": "b633cd1f-f333-4c16-8e85-b43212472666", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.397252] env[65107]: WARNING neutronclient.v2_0.client [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.398009] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.398400] env[65107]: WARNING openstack [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.430666] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 93cd2f0d-6bea-42c2-83b9-085dc94e27ab] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1154.495885] env[65107]: DEBUG nova.network.neutron [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Updated VIF entry in instance network info cache for port 079f6503-3319-4d8a-b94b-843464a762f0. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1154.496297] env[65107]: DEBUG nova.network.neutron [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Updating instance_info_cache with network_info: [{"id": "079f6503-3319-4d8a-b94b-843464a762f0", "address": "fa:16:3e:7b:a0:24", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap079f6503-33", "ovs_interfaceid": "079f6503-3319-4d8a-b94b-843464a762f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.588953] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "interface-0599df85-ad33-4bd5-b234-611db1ef73f0-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.589339] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-0599df85-ad33-4bd5-b234-611db1ef73f0-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.597183] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103650, 'name': CreateVM_Task, 'duration_secs': 1.588955} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.597651] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1154.598285] env[65107]: WARNING neutronclient.v2_0.client [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.598787] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.598863] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.599188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1154.599491] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10c63bef-5808-43a0-bf9b-d643750c9b62 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.605687] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1154.605687] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52361d8d-5f8b-221d-fbaa-026900609d8f" [ 1154.605687] env[65107]: _type = "Task" [ 1154.605687] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.618021] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52361d8d-5f8b-221d-fbaa-026900609d8f, 'name': SearchDatastore_Task} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.618331] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1154.618564] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1154.618849] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.618931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.619124] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1154.619392] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b48bd842-1920-4f47-8c7e-c953589daefd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.628776] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1154.628968] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1154.629880] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f3a83c1-1598-42cb-b3b2-afd4cb7dcb41 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.635981] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1154.635981] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52099793-6eea-4203-4e77-b5a15d58c0db" [ 1154.635981] env[65107]: _type = "Task" [ 1154.635981] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.644230] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52099793-6eea-4203-4e77-b5a15d58c0db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.784688] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103655, 'name': ReconfigVM_Task, 'duration_secs': 0.596934} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.785064] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 0161c1ba-8aff-4f70-a407-caa991e43f15/0161c1ba-8aff-4f70-a407-caa991e43f15.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1154.785701] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-155bf79d-884f-4709-90f6-0cc2cc21df83 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.793157] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1154.793157] env[65107]: value = "task-5103656" [ 1154.793157] env[65107]: _type = "Task" [ 1154.793157] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.801991] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103656, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.856570] env[65107]: DEBUG nova.scheduler.client.report [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1154.873893] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1154.874189] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1154.874343] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1154.874579] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1154.874748] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1154.874913] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1154.875148] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1154.875314] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1154.875477] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1154.876000] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1154.876000] env[65107]: DEBUG nova.virt.hardware [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1154.876362] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-d52a955c-bc3c-41f2-b3ab-db9da613279e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1154.876705] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Instance network_info: |[{"id": "b633cd1f-f333-4c16-8e85-b43212472666", "address": "fa:16:3e:f7:a0:37", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb633cd1f-f3", "ovs_interfaceid": "b633cd1f-f333-4c16-8e85-b43212472666", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1154.877625] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f023be-6cbe-40a3-a192-9802a730a5ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.881094] env[65107]: DEBUG oslo_concurrency.lockutils [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Acquired lock "refresh_cache-d52a955c-bc3c-41f2-b3ab-db9da613279e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.881193] env[65107]: DEBUG nova.network.neutron [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Refreshing network info cache for port b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1154.882365] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:a0:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b633cd1f-f333-4c16-8e85-b43212472666', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1154.889829] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1154.890883] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1154.891594] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c1a0bbe-6fef-4235-a2c3-daeb5b74e8a0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.910876] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a2c214-c292-4939-8c07-174dffd14ab2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.916666] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1154.916666] env[65107]: value = "task-5103657" [ 1154.916666] env[65107]: _type = "Task" [ 1154.916666] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.929306] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:c8:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f9ffee1-f413-4f28-8bc4-3fb2cf299789', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '429e11fc-b198-49e7-bce8-d832c0a6d038', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1154.936871] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1154.937877] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: b0b3e29e-9529-4d6d-b27c-c9d0169c5955] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1154.939736] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1154.944048] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6eacd33b-3cad-4421-9c90-b04d8f37f33a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.959466] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103657, 'name': CreateVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.967042] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1154.967042] env[65107]: value = "task-5103658" [ 1154.967042] env[65107]: _type = "Task" [ 1154.967042] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.976142] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103658, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.999464] env[65107]: DEBUG oslo_concurrency.lockutils [req-0f00748b-a300-4745-b3e4-eabb8e7d9d6b req-52f28051-e3fb-4d3b-8db9-9a7ed7ed0dc4 service nova] Releasing lock "refresh_cache-4f93a52c-9c40-47e5-a34f-b19aea34d7a9" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.098288] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.099060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.099727] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c677d3-f991-4d23-9833-1f9769ad73c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.119352] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c998c640-de28-4abd-83d5-bf4f2186330d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.143614] env[65107]: WARNING neutronclient.v2_0.client [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.149208] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Reconfiguring VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1155.152684] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b480347-3a96-4448-b561-52ff94a8de93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.174401] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52099793-6eea-4203-4e77-b5a15d58c0db, 'name': SearchDatastore_Task, 'duration_secs': 0.009323} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.180033] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1155.180033] env[65107]: value = "task-5103659" [ 1155.180033] env[65107]: _type = "Task" [ 1155.180033] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.180033] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-729feb1c-5405-4b39-838f-94cb565d2296 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.188182] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1155.188182] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5287e955-63e0-2abd-a37d-2048ae76fc8f" [ 1155.188182] env[65107]: _type = "Task" [ 1155.188182] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.194029] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.203324] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5287e955-63e0-2abd-a37d-2048ae76fc8f, 'name': SearchDatastore_Task, 'duration_secs': 0.012078} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.203649] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.203910] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 4f93a52c-9c40-47e5-a34f-b19aea34d7a9/4f93a52c-9c40-47e5-a34f-b19aea34d7a9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1155.204216] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23aebb1b-1184-4428-b02f-a843e2f991ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.216040] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1155.216040] env[65107]: value = "task-5103660" [ 1155.216040] env[65107]: _type = "Task" [ 1155.216040] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.226780] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.310777] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103656, 'name': Rename_Task, 'duration_secs': 0.258664} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.311204] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1155.311531] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3fd273b-386f-4bac-8973-7a94207b4813 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.319369] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1155.319369] env[65107]: value = "task-5103661" [ 1155.319369] env[65107]: _type = "Task" [ 1155.319369] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.329843] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103661, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.365352] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.967s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.366022] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1155.370447] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.195s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.372131] env[65107]: INFO nova.compute.claims [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1155.392511] env[65107]: WARNING neutronclient.v2_0.client [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.393301] env[65107]: WARNING openstack [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.394841] env[65107]: WARNING openstack [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.430918] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103657, 'name': CreateVM_Task, 'duration_secs': 0.424947} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.431917] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1155.432867] env[65107]: WARNING neutronclient.v2_0.client [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.432980] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.433138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.433420] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1155.434025] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5393171a-eaa1-4731-b5ca-c3c054464fd8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.445020] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1155.445020] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526c54d6-fc74-4401-4bce-f6bf329b7099" [ 1155.445020] env[65107]: _type = "Task" [ 1155.445020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.445020] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3f544b3f-703a-49d5-b3e8-16672766f691] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1155.458679] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526c54d6-fc74-4401-4bce-f6bf329b7099, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.481100] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103658, 'name': CreateVM_Task, 'duration_secs': 0.39882} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.481980] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1155.481980] env[65107]: WARNING neutronclient.v2_0.client [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1155.482364] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.592661] env[65107]: WARNING openstack [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.594372] env[65107]: WARNING openstack [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.695434] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.710812] env[65107]: WARNING neutronclient.v2_0.client [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.711563] env[65107]: WARNING openstack [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.711997] env[65107]: WARNING openstack [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.736024] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103660, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.820315] env[65107]: DEBUG nova.network.neutron [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Updated VIF entry in instance network info cache for port b633cd1f-f333-4c16-8e85-b43212472666. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1155.820623] env[65107]: DEBUG nova.network.neutron [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Updating instance_info_cache with network_info: [{"id": "b633cd1f-f333-4c16-8e85-b43212472666", "address": "fa:16:3e:f7:a0:37", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb633cd1f-f3", "ovs_interfaceid": "b633cd1f-f333-4c16-8e85-b43212472666", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1155.832365] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103661, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.879072] env[65107]: DEBUG nova.compute.utils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1155.882997] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1155.883191] env[65107]: DEBUG nova.network.neutron [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1155.883523] env[65107]: WARNING neutronclient.v2_0.client [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.884217] env[65107]: WARNING neutronclient.v2_0.client [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.884598] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.885808] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.935694] env[65107]: DEBUG nova.policy [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5fb11f9a7f72499eb60964784b8b9afe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84acb61dcc764cd79ad2e5b1e2850d16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1155.951887] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f6536493-fb80-4fb1-9ee1-e530734e9adf] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1155.957824] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526c54d6-fc74-4401-4bce-f6bf329b7099, 'name': SearchDatastore_Task, 'duration_secs': 0.063011} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.959236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.959236] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1155.959236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.959236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.959517] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1155.959558] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.959880] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1155.960132] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfd70821-1b10-4766-80ec-1321b4220eb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.964665] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cb3f2b7-6c90-4bc5-8715-a1484e77ce3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.971479] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1155.971479] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e64d2c-cc12-507b-9dd4-a7625aed94a7" [ 1155.971479] env[65107]: _type = "Task" [ 1155.971479] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.976491] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1155.977062] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1155.978311] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-896c1edf-3c01-42cb-ada2-dd314fea6a51 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.985580] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e64d2c-cc12-507b-9dd4-a7625aed94a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.989925] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1155.989925] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5287b6cf-af05-f1ee-581e-51acefcfa756" [ 1155.989925] env[65107]: _type = "Task" [ 1155.989925] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.000937] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5287b6cf-af05-f1ee-581e-51acefcfa756, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.194350] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.231517] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541613} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.232157] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 4f93a52c-9c40-47e5-a34f-b19aea34d7a9/4f93a52c-9c40-47e5-a34f-b19aea34d7a9.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1156.233111] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1156.236768] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78476487-90f3-42b8-be49-8806863967a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.248205] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1156.248205] env[65107]: value = "task-5103662" [ 1156.248205] env[65107]: _type = "Task" [ 1156.248205] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.262115] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103662, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.326820] env[65107]: DEBUG oslo_concurrency.lockutils [req-19a2abe1-41bc-4ca4-a3ac-0f2efcb14b30 req-6531c413-3384-416f-b351-cf923ffb54b1 service nova] Releasing lock "refresh_cache-d52a955c-bc3c-41f2-b3ab-db9da613279e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.333017] env[65107]: DEBUG oslo_vmware.api [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103661, 'name': PowerOnVM_Task, 'duration_secs': 0.536536} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.333964] env[65107]: DEBUG nova.network.neutron [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Successfully created port: 13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1156.336234] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1156.336434] env[65107]: INFO nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Took 8.75 seconds to spawn the instance on the hypervisor. [ 1156.336607] env[65107]: DEBUG nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1156.337625] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482194b3-274d-46cd-bb87-e7336193a992 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.383858] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1156.460870] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 8dbed811-3d4e-4cae-9981-0334801013aa] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1156.488865] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e64d2c-cc12-507b-9dd4-a7625aed94a7, 'name': SearchDatastore_Task, 'duration_secs': 0.012322} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.489123] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.489357] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1156.489570] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.505806] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5287b6cf-af05-f1ee-581e-51acefcfa756, 'name': SearchDatastore_Task, 'duration_secs': 0.010572} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.506936] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e38731be-ed1f-49a6-888e-b3f88931904d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.516057] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1156.516057] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a525ac-5e49-a02e-7cfe-ccc5e5b15e18" [ 1156.516057] env[65107]: _type = "Task" [ 1156.516057] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.525714] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a525ac-5e49-a02e-7cfe-ccc5e5b15e18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.665580] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c99c5e-8851-41aa-a28c-ab03f0705e70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.673553] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50fa0f2-92d7-4a9d-99de-2fa26af6926e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.708824] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b77e9e-6f0e-4fe4-901a-06b70e255303 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.720572] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c6ea69-9e1b-4edd-b3a7-efc319f3dbb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.724749] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.735718] env[65107]: DEBUG nova.compute.provider_tree [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.761343] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103662, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.19081} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.761640] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.762470] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054aa80b-e050-4ec3-93ae-e9c084b6df9f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.788083] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 4f93a52c-9c40-47e5-a34f-b19aea34d7a9/4f93a52c-9c40-47e5-a34f-b19aea34d7a9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.788451] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-418edb0b-dba7-4a10-a273-18653080b30d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.809764] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1156.809764] env[65107]: value = "task-5103663" [ 1156.809764] env[65107]: _type = "Task" [ 1156.809764] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.819706] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103663, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.857142] env[65107]: INFO nova.compute.manager [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Took 18.65 seconds to build instance. [ 1156.964481] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 7373fbbe-52c1-4955-8da9-a5563d71391b] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1157.028519] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a525ac-5e49-a02e-7cfe-ccc5e5b15e18, 'name': SearchDatastore_Task, 'duration_secs': 0.010142} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.028792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.029077] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d52a955c-bc3c-41f2-b3ab-db9da613279e/d52a955c-bc3c-41f2-b3ab-db9da613279e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1157.029384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1157.029573] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1157.029790] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8884a835-78ca-4a10-94f0-1f2c8fdbebfb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.032177] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df2e8142-1aa9-4252-b1cd-3efdcb3a275f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.042564] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1157.042564] env[65107]: value = "task-5103664" [ 1157.042564] env[65107]: _type = "Task" [ 1157.042564] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.047470] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1157.047698] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1157.049141] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24ae9fe2-f130-4558-b1f7-6efc8db757c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.064097] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.065798] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1157.065798] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010ea4-d9b5-fd30-e35f-febe04b71cdf" [ 1157.065798] env[65107]: _type = "Task" [ 1157.065798] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.076280] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52010ea4-d9b5-fd30-e35f-febe04b71cdf, 'name': SearchDatastore_Task, 'duration_secs': 0.010192} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.077023] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f218b6a1-3d01-422c-9ae9-775f8e43222e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.083864] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1157.083864] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526f25c3-0d0f-24f2-0c45-3ca12fabb802" [ 1157.083864] env[65107]: _type = "Task" [ 1157.083864] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.093603] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526f25c3-0d0f-24f2-0c45-3ca12fabb802, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.215836] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.239271] env[65107]: DEBUG nova.scheduler.client.report [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1157.323684] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103663, 'name': ReconfigVM_Task, 'duration_secs': 0.476753} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.324265] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 4f93a52c-9c40-47e5-a34f-b19aea34d7a9/4f93a52c-9c40-47e5-a34f-b19aea34d7a9.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1157.324998] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50d1a970-cd4e-493f-b188-c8e4e15b7801 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.336638] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1157.336638] env[65107]: value = "task-5103665" [ 1157.336638] env[65107]: _type = "Task" [ 1157.336638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.350947] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103665, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.361748] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fb2b3114-2296-4c24-b045-d52b1c98c066 tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.165s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.395065] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1157.426466] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1157.426886] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1157.427032] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1157.427238] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1157.427387] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1157.427566] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1157.427819] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1157.427982] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1157.428213] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1157.428381] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1157.428582] env[65107]: DEBUG nova.virt.hardware [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1157.429616] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368e3367-a92a-4e6e-9ed7-93af48f68fd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.440194] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe838d1-cb20-4354-abfb-dc16f64d63f6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.468559] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: a5ea4075-616a-49ce-aa93-f1db495a022b] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1157.555834] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103664, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.599204] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526f25c3-0d0f-24f2-0c45-3ca12fabb802, 'name': SearchDatastore_Task, 'duration_secs': 0.011023} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.599527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.599899] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1157.600227] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbdd47a8-6478-49a8-8704-54ac461f87cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.609186] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1157.609186] env[65107]: value = "task-5103666" [ 1157.609186] env[65107]: _type = "Task" [ 1157.609186] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.619104] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103666, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.717471] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.746058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.746614] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1157.847999] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103665, 'name': Rename_Task, 'duration_secs': 0.223207} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.848334] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1157.848610] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd103a7c-8067-4e03-b1cc-fb1c872ccba1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.857106] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1157.857106] env[65107]: value = "task-5103667" [ 1157.857106] env[65107]: _type = "Task" [ 1157.857106] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.866687] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.876636] env[65107]: DEBUG nova.compute.manager [req-47e7ac0a-c95a-4d57-9435-861dc7095840 req-33371ad0-2b68-4f2f-bb2b-da9603fd7a8b service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Received event network-vif-plugged-13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1157.876845] env[65107]: DEBUG oslo_concurrency.lockutils [req-47e7ac0a-c95a-4d57-9435-861dc7095840 req-33371ad0-2b68-4f2f-bb2b-da9603fd7a8b service nova] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.877071] env[65107]: DEBUG oslo_concurrency.lockutils [req-47e7ac0a-c95a-4d57-9435-861dc7095840 req-33371ad0-2b68-4f2f-bb2b-da9603fd7a8b service nova] Lock "751ebb31-1a27-4630-9213-4312a6fa9298-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.877244] env[65107]: DEBUG oslo_concurrency.lockutils [req-47e7ac0a-c95a-4d57-9435-861dc7095840 req-33371ad0-2b68-4f2f-bb2b-da9603fd7a8b service nova] Lock "751ebb31-1a27-4630-9213-4312a6fa9298-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.877414] env[65107]: DEBUG nova.compute.manager [req-47e7ac0a-c95a-4d57-9435-861dc7095840 req-33371ad0-2b68-4f2f-bb2b-da9603fd7a8b service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] No waiting events found dispatching network-vif-plugged-13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1157.877571] env[65107]: WARNING nova.compute.manager [req-47e7ac0a-c95a-4d57-9435-861dc7095840 req-33371ad0-2b68-4f2f-bb2b-da9603fd7a8b service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Received unexpected event network-vif-plugged-13b14810-c694-4e47-b090-8c3bf39ad510 for instance with vm_state building and task_state spawning. [ 1157.924708] env[65107]: DEBUG nova.network.neutron [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Successfully updated port: 13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1157.973081] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 04a508b4-4430-411c-8573-726065b558a1] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1158.056450] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592094} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.056752] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d52a955c-bc3c-41f2-b3ab-db9da613279e/d52a955c-bc3c-41f2-b3ab-db9da613279e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1158.057009] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1158.057341] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86d51fcf-7985-41ce-b983-8cc52adcd510 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.070074] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1158.070074] env[65107]: value = "task-5103668" [ 1158.070074] env[65107]: _type = "Task" [ 1158.070074] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.083718] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103668, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.123363] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103666, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.153558] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "0161c1ba-8aff-4f70-a407-caa991e43f15" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.153838] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.154084] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "0161c1ba-8aff-4f70-a407-caa991e43f15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.154293] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.154463] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.157380] env[65107]: INFO nova.compute.manager [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Terminating instance [ 1158.225531] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.252511] env[65107]: DEBUG nova.compute.utils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1158.254218] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1158.254461] env[65107]: DEBUG nova.network.neutron [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1158.254853] env[65107]: WARNING neutronclient.v2_0.client [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.255255] env[65107]: WARNING neutronclient.v2_0.client [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.255944] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.256360] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.321316] env[65107]: DEBUG nova.policy [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ded1c3ef07d4e78886dc924e9099363', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '782397ddd6cb40298dcf7f0da7353d67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1158.365179] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.365458] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.372862] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103667, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.427295] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.427499] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.427687] env[65107]: DEBUG nova.network.neutron [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1158.476406] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d67468c6-9aaf-49c4-afe4-5d9856c4af7a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1158.585125] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103668, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127412} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.585513] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1158.586759] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24219ad-b7fe-4037-8440-b3d72773298b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.616319] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] d52a955c-bc3c-41f2-b3ab-db9da613279e/d52a955c-bc3c-41f2-b3ab-db9da613279e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.616788] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-853b85fb-c649-4118-a0ee-a755da0654d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.645725] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103666, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662758} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.647316] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1158.647672] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1158.647964] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1158.647964] env[65107]: value = "task-5103669" [ 1158.647964] env[65107]: _type = "Task" [ 1158.647964] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.648214] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87eb9bd4-6745-41d4-a1fe-4962fe541a17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.661711] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103669, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.664575] env[65107]: DEBUG nova.compute.manager [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1158.664646] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1158.665436] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1158.665436] env[65107]: value = "task-5103670" [ 1158.665436] env[65107]: _type = "Task" [ 1158.665436] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.666240] env[65107]: DEBUG nova.network.neutron [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Successfully created port: ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1158.669888] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0f54fa-0146-456b-b5d4-7fc53da84c93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.688376] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1158.692123] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e85fec4-3926-403d-aa2a-53473e1a86e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.694347] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103670, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.700621] env[65107]: DEBUG oslo_vmware.api [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1158.700621] env[65107]: value = "task-5103671" [ 1158.700621] env[65107]: _type = "Task" [ 1158.700621] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.712227] env[65107]: DEBUG oslo_vmware.api [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.723689] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.765605] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1158.868672] env[65107]: DEBUG oslo_vmware.api [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103667, 'name': PowerOnVM_Task, 'duration_secs': 0.835613} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.868988] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1158.869217] env[65107]: INFO nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Took 8.84 seconds to spawn the instance on the hypervisor. [ 1158.869399] env[65107]: DEBUG nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1158.870219] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13b08d4-8245-4607-b12e-adf9f51a2d6f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.874245] env[65107]: DEBUG nova.compute.utils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1158.931145] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.932065] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.979937] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 00574b77-dad6-4f0a-bbcc-20a2b4a5df65] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1158.985409] env[65107]: DEBUG nova.network.neutron [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1159.006736] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.007183] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.108698] env[65107]: WARNING neutronclient.v2_0.client [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.109376] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.109731] env[65107]: WARNING openstack [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.166033] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103669, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.185392] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103670, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.173457} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.185772] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1159.186967] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493b527f-019e-409b-aef5-c2991fafd582 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.213677] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.226332] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b925c2d1-c0ed-4924-886d-d9709de67798 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.255010] env[65107]: DEBUG oslo_vmware.api [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103671, 'name': PowerOffVM_Task, 'duration_secs': 0.502456} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.259600] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1159.259938] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1159.260350] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.260765] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1159.260765] env[65107]: value = "task-5103672" [ 1159.260765] env[65107]: _type = "Task" [ 1159.260765] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.261081] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-603cc41d-4b43-4bde-9f88-95a94e5a9f12 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.278530] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103672, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.296241] env[65107]: DEBUG nova.network.neutron [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [{"id": "13b14810-c694-4e47-b090-8c3bf39ad510", "address": "fa:16:3e:e7:23:0e", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13b14810-c6", "ovs_interfaceid": "13b14810-c694-4e47-b090-8c3bf39ad510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.384179] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.019s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.392988] env[65107]: INFO nova.compute.manager [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Took 19.33 seconds to build instance. [ 1159.482960] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: b4f7d003-59fb-4ef4-b5f6-234d5154d198] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1159.622168] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1159.622546] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1159.622845] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleting the datastore file [datastore2] 0161c1ba-8aff-4f70-a407-caa991e43f15 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1159.623209] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66ed36ce-cdd8-4526-8de7-074091fb16cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.632496] env[65107]: DEBUG oslo_vmware.api [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for the task: (returnval){ [ 1159.632496] env[65107]: value = "task-5103674" [ 1159.632496] env[65107]: _type = "Task" [ 1159.632496] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.642792] env[65107]: DEBUG oslo_vmware.api [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.661989] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103669, 'name': ReconfigVM_Task, 'duration_secs': 0.743356} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.662307] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Reconfigured VM instance instance-00000074 to attach disk [datastore2] d52a955c-bc3c-41f2-b3ab-db9da613279e/d52a955c-bc3c-41f2-b3ab-db9da613279e.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1159.663022] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ebac9c4-3f4c-4195-8ea8-377b9ca25b40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.671583] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1159.671583] env[65107]: value = "task-5103675" [ 1159.671583] env[65107]: _type = "Task" [ 1159.671583] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.682298] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103675, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.736953] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.775347] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.780817] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1159.799021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.799713] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Instance network_info: |[{"id": "13b14810-c694-4e47-b090-8c3bf39ad510", "address": "fa:16:3e:e7:23:0e", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13b14810-c6", "ovs_interfaceid": "13b14810-c694-4e47-b090-8c3bf39ad510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1159.800270] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:23:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13b14810-c694-4e47-b090-8c3bf39ad510', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1159.809457] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Creating folder: Project (84acb61dcc764cd79ad2e5b1e2850d16). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1159.811872] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06dd0853-d0c7-48bb-a201-6c13ac06e29d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.821604] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1159.821867] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1159.822039] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1159.822228] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1159.822373] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1159.822566] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1159.822786] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1159.822946] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1159.823130] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1159.823295] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1159.823464] env[65107]: DEBUG nova.virt.hardware [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1159.824573] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459bfda5-3c75-47f1-8658-4a31df908c3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.829086] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Created folder: Project (84acb61dcc764cd79ad2e5b1e2850d16) in parent group-v992574. [ 1159.829301] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Creating folder: Instances. Parent ref: group-v992883. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1159.830065] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e5cde09-87c3-4268-8e14-23112f8c0370 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.836664] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20493c3-ba9b-4cc3-be65-e1fa7b17cf61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.844154] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Created folder: Instances in parent group-v992883. [ 1159.844427] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1159.844660] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1159.844913] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd8c1009-9346-4659-8e08-1c9544761f55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.874252] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1159.874252] env[65107]: value = "task-5103678" [ 1159.874252] env[65107]: _type = "Task" [ 1159.874252] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.882899] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103678, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.895678] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c8729e8a-70cd-46a4-a55a-1b8cba57370d tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.846s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.986577] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: a3162257-ac6b-4468-a573-3611bba69c46] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1160.099038] env[65107]: DEBUG nova.compute.manager [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Received event network-changed-13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1160.099633] env[65107]: DEBUG nova.compute.manager [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Refreshing instance network info cache due to event network-changed-13b14810-c694-4e47-b090-8c3bf39ad510. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1160.099708] env[65107]: DEBUG oslo_concurrency.lockutils [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Acquiring lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.100888] env[65107]: DEBUG oslo_concurrency.lockutils [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Acquired lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.100888] env[65107]: DEBUG nova.network.neutron [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Refreshing network info cache for port 13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1160.147367] env[65107]: DEBUG oslo_vmware.api [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Task: {'id': task-5103674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175014} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.147367] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1160.147367] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1160.147367] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1160.147777] env[65107]: INFO nova.compute.manager [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Took 1.48 seconds to destroy the instance on the hypervisor. [ 1160.147811] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1160.148398] env[65107]: DEBUG nova.compute.manager [-] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1160.148398] env[65107]: DEBUG nova.network.neutron [-] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1160.148398] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.148875] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.149168] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.184822] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103675, 'name': Rename_Task, 'duration_secs': 0.320734} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.185091] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1160.185330] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d09a9cb-2422-4411-97e9-dacdc19be508 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.193638] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1160.193638] env[65107]: value = "task-5103679" [ 1160.193638] env[65107]: _type = "Task" [ 1160.193638] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.201110] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.210502] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.240982] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.242541] env[65107]: DEBUG nova.network.neutron [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Successfully updated port: ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1160.278400] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103672, 'name': ReconfigVM_Task, 'duration_secs': 0.810153} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.279545] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Reconfigured VM instance instance-00000069 to attach disk [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896/d7b4bc1e-c94e-4654-9345-2a8aa945a896.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.280253] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2b12c0a-2285-4bc1-ad65-523fce0559d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.293252] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1160.293252] env[65107]: value = "task-5103680" [ 1160.293252] env[65107]: _type = "Task" [ 1160.293252] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.308063] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103680, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.386815] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103678, 'name': CreateVM_Task, 'duration_secs': 0.404196} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.387216] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1160.387924] env[65107]: WARNING neutronclient.v2_0.client [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.388525] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.388781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.389277] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1160.389687] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51c5283f-efda-497f-be2e-f36544863a5c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.397028] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1160.397028] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528d3fd8-5537-6bea-381a-9d2e628a9ae7" [ 1160.397028] env[65107]: _type = "Task" [ 1160.397028] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.407799] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528d3fd8-5537-6bea-381a-9d2e628a9ae7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.442668] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.442946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1160.443219] env[65107]: INFO nova.compute.manager [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Attaching volume f451fb1b-2805-4784-8cf8-a2cce68d18d1 to /dev/sdb [ 1160.492019] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71eb9b59-ba7e-45ad-b64a-cad257a34d17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.493167] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 8cddc4d1-c94c-4f2b-9c4b-3125e4638050] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1160.501324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a18206e-7627-4c87-b186-d98796e18a74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.524279] env[65107]: DEBUG nova.virt.block_device [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating existing volume attachment record: bfc420fb-8cb9-4057-b11b-411b61000f97 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1160.591124] env[65107]: DEBUG nova.compute.manager [req-51106937-c957-43c6-8396-0a732f505153 req-4238ddad-1d66-4201-96a4-bbed1928f063 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Received event network-vif-deleted-3e149eec-773a-449a-8c36-4dd31d98e502 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1160.591353] env[65107]: INFO nova.compute.manager [req-51106937-c957-43c6-8396-0a732f505153 req-4238ddad-1d66-4201-96a4-bbed1928f063 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Neutron deleted interface 3e149eec-773a-449a-8c36-4dd31d98e502; detaching it from the instance and deleting it from the info cache [ 1160.591539] env[65107]: DEBUG nova.network.neutron [req-51106937-c957-43c6-8396-0a732f505153 req-4238ddad-1d66-4201-96a4-bbed1928f063 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1160.603704] env[65107]: WARNING neutronclient.v2_0.client [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.604378] env[65107]: WARNING openstack [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.604763] env[65107]: WARNING openstack [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.709648] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103679, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.728057] env[65107]: WARNING openstack [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.728477] env[65107]: WARNING openstack [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.748841] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "refresh_cache-7e72891b-4c20-4889-80c9-8bcb15373a2c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.749155] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "refresh_cache-7e72891b-4c20-4889-80c9-8bcb15373a2c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.749244] env[65107]: DEBUG nova.network.neutron [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1160.750834] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.805703] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103680, 'name': Rename_Task, 'duration_secs': 0.183933} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.807268] env[65107]: WARNING neutronclient.v2_0.client [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.808274] env[65107]: WARNING openstack [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.808499] env[65107]: WARNING openstack [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.817035] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1160.817720] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7172155c-20f8-494b-8f72-44465d49b6c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.826886] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1160.826886] env[65107]: value = "task-5103684" [ 1160.826886] env[65107]: _type = "Task" [ 1160.826886] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.838579] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.909157] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]528d3fd8-5537-6bea-381a-9d2e628a9ae7, 'name': SearchDatastore_Task, 'duration_secs': 0.011805} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.909623] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1160.909866] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.910176] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.910388] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.910618] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.910998] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cbc3a1d-6cfa-4b22-bd72-040a0f2870ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.916766] env[65107]: DEBUG nova.network.neutron [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updated VIF entry in instance network info cache for port 13b14810-c694-4e47-b090-8c3bf39ad510. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1160.917214] env[65107]: DEBUG nova.network.neutron [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [{"id": "13b14810-c694-4e47-b090-8c3bf39ad510", "address": "fa:16:3e:e7:23:0e", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13b14810-c6", "ovs_interfaceid": "13b14810-c694-4e47-b090-8c3bf39ad510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1160.922255] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.922546] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1160.923477] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-608dd920-3737-426b-9693-948fe60dac2c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.930787] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1160.930787] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52746f9d-c73b-597d-db9d-ad183dcc88fa" [ 1160.930787] env[65107]: _type = "Task" [ 1160.930787] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.943829] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52746f9d-c73b-597d-db9d-ad183dcc88fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.993363] env[65107]: DEBUG nova.network.neutron [-] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1160.996424] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 59ff1aa6-8d0d-4885-90d3-33d623dca8bb] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1161.095178] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d75f6b51-4c9a-4d57-ae2a-57cca1648802 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.106649] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40568f5b-c373-4db4-b802-f81f27f47287 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.145923] env[65107]: DEBUG nova.compute.manager [req-51106937-c957-43c6-8396-0a732f505153 req-4238ddad-1d66-4201-96a4-bbed1928f063 service nova] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Detach interface failed, port_id=3e149eec-773a-449a-8c36-4dd31d98e502, reason: Instance 0161c1ba-8aff-4f70-a407-caa991e43f15 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1161.206259] env[65107]: DEBUG oslo_vmware.api [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103679, 'name': PowerOnVM_Task, 'duration_secs': 0.69295} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.206542] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1161.206829] env[65107]: INFO nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Took 8.74 seconds to spawn the instance on the hypervisor. [ 1161.207036] env[65107]: DEBUG nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1161.207831] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07613db-e7e9-46cd-bb79-513e1b4e7dff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.246234] env[65107]: DEBUG oslo_vmware.api [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103659, 'name': ReconfigVM_Task, 'duration_secs': 5.820172} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.246531] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.246804] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Reconfigured VM to detach interface {{(pid=65107) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1161.247281] env[65107]: WARNING neutronclient.v2_0.client [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.247593] env[65107]: WARNING neutronclient.v2_0.client [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.248222] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.248601] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.256585] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.257042] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.300530] env[65107]: WARNING neutronclient.v2_0.client [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.304468] env[65107]: DEBUG nova.network.neutron [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1161.333363] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.333770] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.352714] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103684, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.413755] env[65107]: WARNING neutronclient.v2_0.client [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.414636] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.415208] env[65107]: WARNING openstack [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.424161] env[65107]: DEBUG oslo_concurrency.lockutils [req-fb22ba73-435a-43ea-9c4a-581ef476ac88 req-39a7f136-5823-4eb3-96b5-e0949a34fbd1 service nova] Releasing lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.442220] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52746f9d-c73b-597d-db9d-ad183dcc88fa, 'name': SearchDatastore_Task, 'duration_secs': 0.012865} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.443133] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3850f82c-c4c4-478b-be8e-becfaf29710d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.450957] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1161.450957] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f71cd9-12bf-3901-8ce7-80e23c5dff39" [ 1161.450957] env[65107]: _type = "Task" [ 1161.450957] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.463146] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f71cd9-12bf-3901-8ce7-80e23c5dff39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.491339] env[65107]: DEBUG nova.compute.manager [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1161.493090] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37444821-c3fb-495f-8853-ed1f31fe4fe9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.496995] env[65107]: INFO nova.compute.manager [-] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Took 1.35 seconds to deallocate network for instance. [ 1161.499364] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 6d5eb77c-c6e8-4ef1-a945-b3485aedd488] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1161.541497] env[65107]: DEBUG nova.network.neutron [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Updating instance_info_cache with network_info: [{"id": "ebe5fc14-2b6a-40f0-80c2-2f843e929082", "address": "fa:16:3e:76:e3:bf", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebe5fc14-2b", "ovs_interfaceid": "ebe5fc14-2b6a-40f0-80c2-2f843e929082", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1161.728125] env[65107]: INFO nova.compute.manager [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Took 19.72 seconds to build instance. [ 1161.841665] env[65107]: DEBUG oslo_vmware.api [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103684, 'name': PowerOnVM_Task, 'duration_secs': 0.644615} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.841833] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1161.841992] env[65107]: DEBUG nova.compute.manager [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1161.843189] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fa9cb7-b65e-47ca-b4e4-24de1552391c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.962844] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f71cd9-12bf-3901-8ce7-80e23c5dff39, 'name': SearchDatastore_Task, 'duration_secs': 0.01306} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.963264] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.963357] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 751ebb31-1a27-4630-9213-4312a6fa9298/751ebb31-1a27-4630-9213-4312a6fa9298.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1161.963624] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b997035-5353-47c1-b1d1-2a78b320cad1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.972791] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1161.972791] env[65107]: value = "task-5103685" [ 1161.972791] env[65107]: _type = "Task" [ 1161.972791] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.982854] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.010736] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 6be8d8aa-6ac2-415d-8a4a-b16de3f0ca21] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1162.014577] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.014964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.015300] env[65107]: DEBUG nova.objects.instance [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lazy-loading 'resources' on Instance uuid 0161c1ba-8aff-4f70-a407-caa991e43f15 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.017524] env[65107]: INFO nova.compute.manager [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] instance snapshotting [ 1162.020601] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ec9622-f281-4dbb-a022-c9c6f2257286 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.046187] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "refresh_cache-7e72891b-4c20-4889-80c9-8bcb15373a2c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1162.046755] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Instance network_info: |[{"id": "ebe5fc14-2b6a-40f0-80c2-2f843e929082", "address": "fa:16:3e:76:e3:bf", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebe5fc14-2b", "ovs_interfaceid": "ebe5fc14-2b6a-40f0-80c2-2f843e929082", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1162.049586] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:e3:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '17c839f5-4de0-449c-9a24-4e0e2fca37ca', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebe5fc14-2b6a-40f0-80c2-2f843e929082', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1162.058174] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1162.058951] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653b3997-5711-4ea8-be39-22cac04e5ca9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.065246] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1162.065825] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4968c7b-1cde-438e-ba7c-4245a2d3e8da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.091989] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1162.091989] env[65107]: value = "task-5103686" [ 1162.091989] env[65107]: _type = "Task" [ 1162.091989] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.103136] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103686, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.177187] env[65107]: DEBUG nova.compute.manager [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Received event network-vif-plugged-ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1162.177962] env[65107]: DEBUG oslo_concurrency.lockutils [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Acquiring lock "7e72891b-4c20-4889-80c9-8bcb15373a2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.178450] env[65107]: DEBUG oslo_concurrency.lockutils [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.178450] env[65107]: DEBUG oslo_concurrency.lockutils [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.178851] env[65107]: DEBUG nova.compute.manager [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] No waiting events found dispatching network-vif-plugged-ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1162.179117] env[65107]: WARNING nova.compute.manager [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Received unexpected event network-vif-plugged-ebe5fc14-2b6a-40f0-80c2-2f843e929082 for instance with vm_state building and task_state spawning. [ 1162.179450] env[65107]: DEBUG nova.compute.manager [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Received event network-changed-ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1162.179772] env[65107]: DEBUG nova.compute.manager [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Refreshing instance network info cache due to event network-changed-ebe5fc14-2b6a-40f0-80c2-2f843e929082. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1162.180141] env[65107]: DEBUG oslo_concurrency.lockutils [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Acquiring lock "refresh_cache-7e72891b-4c20-4889-80c9-8bcb15373a2c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.180367] env[65107]: DEBUG oslo_concurrency.lockutils [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Acquired lock "refresh_cache-7e72891b-4c20-4889-80c9-8bcb15373a2c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.180614] env[65107]: DEBUG nova.network.neutron [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Refreshing network info cache for port ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1162.231128] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0e9c29b1-ae88-449d-b1c3-ee9c0dbd5808 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.242s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.263544] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.263897] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.264159] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.264461] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.264755] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.268024] env[65107]: INFO nova.compute.manager [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Terminating instance [ 1162.364139] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.485839] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103685, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.518482] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 13c96e2f-46de-46e0-a505-60dfec9e95ca] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1162.566626] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "d52a955c-bc3c-41f2-b3ab-db9da613279e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.566895] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.567128] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "d52a955c-bc3c-41f2-b3ab-db9da613279e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.567315] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.567494] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.569812] env[65107]: INFO nova.compute.manager [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Terminating instance [ 1162.593761] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1162.594459] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-46a46cb5-8167-492f-8f01-48bc6bc91912 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.608174] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103686, 'name': CreateVM_Task, 'duration_secs': 0.448567} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.609458] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1162.609813] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1162.609813] env[65107]: value = "task-5103687" [ 1162.609813] env[65107]: _type = "Task" [ 1162.609813] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.610280] env[65107]: WARNING neutronclient.v2_0.client [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.610640] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.610782] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.611127] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1162.611476] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79602bc0-d9c5-4755-8f14-42057dde2312 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.626237] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103687, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.626808] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1162.626808] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e55481-e80a-48d5-346e-8386e377a55b" [ 1162.626808] env[65107]: _type = "Task" [ 1162.626808] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.637915] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e55481-e80a-48d5-346e-8386e377a55b, 'name': SearchDatastore_Task, 'duration_secs': 0.011402} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.638254] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1162.638483] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1162.638805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.638985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.639251] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1162.639550] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb145d71-0cee-4361-b003-7084594be047 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.654366] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1162.654589] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1162.655383] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b6d68cf-473d-40f0-8444-00280812bc7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.664253] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1162.664253] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527c4a95-d3e5-5905-ecef-31987d333957" [ 1162.664253] env[65107]: _type = "Task" [ 1162.664253] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.673686] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.673858] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquired lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.674051] env[65107]: DEBUG nova.network.neutron [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1162.679304] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527c4a95-d3e5-5905-ecef-31987d333957, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.685662] env[65107]: WARNING neutronclient.v2_0.client [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.686377] env[65107]: WARNING openstack [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.686741] env[65107]: WARNING openstack [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.773582] env[65107]: DEBUG nova.compute.manager [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1162.773833] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.777540] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccfbc41-09fc-4e02-8407-8572a6bde4d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.787841] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1162.792811] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8433c47-a880-4a29-a1e9-c14307e1277e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.796801] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4f554f-ddea-485d-abc3-41f24bf781ae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.806874] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b504fa-f73a-47b0-81ea-81a05e168413 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.811492] env[65107]: DEBUG oslo_vmware.api [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1162.811492] env[65107]: value = "task-5103688" [ 1162.811492] env[65107]: _type = "Task" [ 1162.811492] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.858056] env[65107]: WARNING openstack [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.858735] env[65107]: WARNING openstack [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.872362] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771cc868-08da-45a0-b5af-e55714edf9f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.880183] env[65107]: DEBUG oslo_vmware.api [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.886611] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcb8247-6a27-421e-ad06-a48a182e47ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.903384] env[65107]: DEBUG nova.compute.provider_tree [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.945325] env[65107]: WARNING neutronclient.v2_0.client [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.945830] env[65107]: WARNING openstack [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.946238] env[65107]: WARNING openstack [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.986366] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56572} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.986878] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 751ebb31-1a27-4630-9213-4312a6fa9298/751ebb31-1a27-4630-9213-4312a6fa9298.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1162.986878] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.987571] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-937f35bf-1704-43de-a73f-4ac385ea3e71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.000702] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1163.000702] env[65107]: value = "task-5103690" [ 1163.000702] env[65107]: _type = "Task" [ 1163.000702] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.014151] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.025037] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f4b1352b-1c55-4987-a298-69431c4e565d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1163.062135] env[65107]: DEBUG nova.network.neutron [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Updated VIF entry in instance network info cache for port ebe5fc14-2b6a-40f0-80c2-2f843e929082. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1163.062585] env[65107]: DEBUG nova.network.neutron [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Updating instance_info_cache with network_info: [{"id": "ebe5fc14-2b6a-40f0-80c2-2f843e929082", "address": "fa:16:3e:76:e3:bf", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebe5fc14-2b", "ovs_interfaceid": "ebe5fc14-2b6a-40f0-80c2-2f843e929082", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.076624] env[65107]: DEBUG nova.compute.manager [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1163.076918] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1163.078093] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d8ab6c-69ff-4e28-bb36-458e233729aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.086464] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1163.086735] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7839fbfe-7785-410d-ada3-b4bde01828ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.093997] env[65107]: DEBUG oslo_vmware.api [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1163.093997] env[65107]: value = "task-5103691" [ 1163.093997] env[65107]: _type = "Task" [ 1163.093997] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.105517] env[65107]: DEBUG oslo_vmware.api [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.122762] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103687, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.179517] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527c4a95-d3e5-5905-ecef-31987d333957, 'name': SearchDatastore_Task, 'duration_secs': 0.010798} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.180244] env[65107]: WARNING neutronclient.v2_0.client [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.181202] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.181639] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.193491] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c879ed1b-7be3-436f-a516-444b276ee6a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.202723] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1163.202723] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a0a855-9241-6e13-7923-5caafc2ad9f5" [ 1163.202723] env[65107]: _type = "Task" [ 1163.202723] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.216686] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52a0a855-9241-6e13-7923-5caafc2ad9f5, 'name': SearchDatastore_Task, 'duration_secs': 0.01156} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.216839] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.217039] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 7e72891b-4c20-4889-80c9-8bcb15373a2c/7e72891b-4c20-4889-80c9-8bcb15373a2c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1163.217351] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b2a7bd4-2ae6-4e91-9d00-d2ad0b6fcb9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.228658] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1163.228658] env[65107]: value = "task-5103692" [ 1163.228658] env[65107]: _type = "Task" [ 1163.228658] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.239044] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103692, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.326475] env[65107]: DEBUG oslo_vmware.api [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103688, 'name': PowerOffVM_Task, 'duration_secs': 0.238794} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.328883] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.329285] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.340123] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1163.340358] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1163.341427] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4e6fdb4-e689-4267-a163-ddf2071f1730 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.411593] env[65107]: DEBUG nova.scheduler.client.report [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1163.433652] env[65107]: WARNING neutronclient.v2_0.client [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.434371] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.434877] env[65107]: WARNING openstack [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.516622] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082028} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.517023] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1163.517981] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a41aa85-15b6-428b-800f-0c8a685257c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.527217] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.527426] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.527617] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleting the datastore file [datastore2] 0599df85-ad33-4bd5-b234-611db1ef73f0 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.539661] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15164d17-ca83-4328-823f-0bfe623a6e4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.542301] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 58345821-536a-46ad-af55-92de7552e924] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1163.554678] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] 751ebb31-1a27-4630-9213-4312a6fa9298/751ebb31-1a27-4630-9213-4312a6fa9298.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1163.555408] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2675bc8-c64f-4678-bb29-ce58d14c398f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.574114] env[65107]: INFO nova.network.neutron [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Port 12711066-1ddb-4f73-aaf9-f2eb3d8c4515 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1163.574592] env[65107]: DEBUG nova.network.neutron [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [{"id": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "address": "fa:16:3e:e8:eb:53", "network": {"id": "40a76aa5-ac46-4067-a98d-cd60410bf479", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-835657332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "152868ee0723449baac10f6a8ceeb738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd04b7a1c-d3", "ovs_interfaceid": "d04b7a1c-d3b0-4211-b44b-2e39791f7497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.576791] env[65107]: DEBUG oslo_concurrency.lockutils [req-4af702c7-2320-4ebb-ba61-83245cfb9302 req-6adcd2e1-23eb-44a3-b2b1-544eeeb2f7c2 service nova] Releasing lock "refresh_cache-7e72891b-4c20-4889-80c9-8bcb15373a2c" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.583707] env[65107]: DEBUG oslo_vmware.api [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1163.583707] env[65107]: value = "task-5103694" [ 1163.583707] env[65107]: _type = "Task" [ 1163.583707] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.585705] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1163.585705] env[65107]: value = "task-5103695" [ 1163.585705] env[65107]: _type = "Task" [ 1163.585705] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.606672] env[65107]: DEBUG oslo_vmware.api [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103694, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.612903] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.623103] env[65107]: DEBUG oslo_vmware.api [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103691, 'name': PowerOffVM_Task, 'duration_secs': 0.425291} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.627267] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1163.627582] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1163.628426] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103687, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.628733] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91976ec6-2a42-4faf-be83-a94ec2ae97e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.736997] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.737302] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.737496] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore2] d52a955c-bc3c-41f2-b3ab-db9da613279e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.737942] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95552dda-dccc-4bcb-b692-868172bfb0c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.743723] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103692, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.751355] env[65107]: DEBUG oslo_vmware.api [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1163.751355] env[65107]: value = "task-5103697" [ 1163.751355] env[65107]: _type = "Task" [ 1163.751355] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.762526] env[65107]: DEBUG oslo_vmware.api [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103697, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.918781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.921277] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.557s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.921503] env[65107]: DEBUG nova.objects.instance [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1163.945170] env[65107]: INFO nova.scheduler.client.report [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Deleted allocations for instance 0161c1ba-8aff-4f70-a407-caa991e43f15 [ 1164.057052] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: c5fb4b1c-5540-4645-b996-f397729beb23] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1164.078711] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Releasing lock "refresh_cache-0599df85-ad33-4bd5-b234-611db1ef73f0" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1164.106397] env[65107]: DEBUG oslo_vmware.api [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103694, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.414514} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.107132] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.107434] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.107649] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.107873] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.108109] env[65107]: INFO nova.compute.manager [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Took 1.33 seconds to destroy the instance on the hypervisor. [ 1164.108431] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1164.108669] env[65107]: DEBUG nova.compute.manager [-] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1164.108782] env[65107]: DEBUG nova.network.neutron [-] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1164.109122] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.109767] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.110091] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.131084] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103687, 'name': CreateSnapshot_Task, 'duration_secs': 1.144255} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.131413] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1164.132508] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06cd664-ee56-46a7-b4cd-1cb9dcf19fd8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.178706] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.242299] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103692, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550597} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.242769] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 7e72891b-4c20-4889-80c9-8bcb15373a2c/7e72891b-4c20-4889-80c9-8bcb15373a2c.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1164.243102] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1164.243509] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d0834af-f3cc-46d9-8744-662c1393fe4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.252403] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1164.252403] env[65107]: value = "task-5103698" [ 1164.252403] env[65107]: _type = "Task" [ 1164.252403] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.267820] env[65107]: DEBUG oslo_vmware.api [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103697, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255802} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.271544] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.271831] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.272109] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.272366] env[65107]: INFO nova.compute.manager [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1164.272639] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1164.272871] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103698, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.277805] env[65107]: DEBUG nova.compute.manager [-] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1164.277947] env[65107]: DEBUG nova.network.neutron [-] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1164.278205] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.278768] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.279061] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.326721] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.455012] env[65107]: DEBUG oslo_concurrency.lockutils [None req-478f76ca-e7b3-40e8-9f0f-479ffaeef6fb tempest-ServerDiskConfigTestJSON-1771527805 tempest-ServerDiskConfigTestJSON-1771527805-project-member] Lock "0161c1ba-8aff-4f70-a407-caa991e43f15" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.301s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.560527] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: e12cb11f-7d6f-47d6-a31e-bf2654174d38] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1164.577032] env[65107]: DEBUG nova.compute.manager [req-e62d44b4-e47e-4755-bb34-779ea52f7190 req-0f992fe9-0baa-4af7-93a3-701776a2d3fa service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Received event network-vif-deleted-b633cd1f-f333-4c16-8e85-b43212472666 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1164.577032] env[65107]: INFO nova.compute.manager [req-e62d44b4-e47e-4755-bb34-779ea52f7190 req-0f992fe9-0baa-4af7-93a3-701776a2d3fa service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Neutron deleted interface b633cd1f-f333-4c16-8e85-b43212472666; detaching it from the instance and deleting it from the info cache [ 1164.577032] env[65107]: DEBUG nova.network.neutron [req-e62d44b4-e47e-4755-bb34-779ea52f7190 req-0f992fe9-0baa-4af7-93a3-701776a2d3fa service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1164.583599] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c9434f04-48b4-41e3-bec5-de1d345df200 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "interface-0599df85-ad33-4bd5-b234-611db1ef73f0-12711066-1ddb-4f73-aaf9-f2eb3d8c4515" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.994s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.606029] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103695, 'name': ReconfigVM_Task, 'duration_secs': 0.583318} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.606371] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Reconfigured VM instance instance-00000075 to attach disk [datastore1] 751ebb31-1a27-4630-9213-4312a6fa9298/751ebb31-1a27-4630-9213-4312a6fa9298.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1164.607420] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6ca1040-95e1-4878-a7e0-e6e7249e410c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.616063] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1164.616063] env[65107]: value = "task-5103699" [ 1164.616063] env[65107]: _type = "Task" [ 1164.616063] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.626588] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103699, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.658506] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1164.658886] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b09bdafc-2bd0-4fc7-af61-842793f2bed6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.669859] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1164.669859] env[65107]: value = "task-5103700" [ 1164.669859] env[65107]: _type = "Task" [ 1164.669859] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.680593] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103700, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.768044] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103698, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073913} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.768374] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1164.769280] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c61bda-d408-4cd2-8e40-0a4f9c95fe45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.795208] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 7e72891b-4c20-4889-80c9-8bcb15373a2c/7e72891b-4c20-4889-80c9-8bcb15373a2c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.795844] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5dd86833-d35e-40ae-82c6-b4eebc8228fa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.820374] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1164.820374] env[65107]: value = "task-5103701" [ 1164.820374] env[65107]: _type = "Task" [ 1164.820374] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.834289] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103701, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.932782] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0b73e101-1ac9-48d3-9fb3-4735566ad0a5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.050934] env[65107]: DEBUG nova.network.neutron [-] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1165.065195] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 52bd7ecb-4f53-443e-8916-3ff50dfab2b6] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1165.082100] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7b931a35-c765-451e-b41f-a5f51a340dc0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.089588] env[65107]: DEBUG nova.compute.manager [req-741a4baa-3c0e-46de-9ed2-1a92c7651437 req-e3ff3f7e-18cb-46e7-96fb-362fad13fd67 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Received event network-vif-deleted-d04b7a1c-d3b0-4211-b44b-2e39791f7497 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1165.089670] env[65107]: INFO nova.compute.manager [req-741a4baa-3c0e-46de-9ed2-1a92c7651437 req-e3ff3f7e-18cb-46e7-96fb-362fad13fd67 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Neutron deleted interface d04b7a1c-d3b0-4211-b44b-2e39791f7497; detaching it from the instance and deleting it from the info cache [ 1165.094434] env[65107]: DEBUG nova.network.neutron [req-741a4baa-3c0e-46de-9ed2-1a92c7651437 req-e3ff3f7e-18cb-46e7-96fb-362fad13fd67 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1165.105730] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d50346-c3c1-4914-8bb1-908aa3848d5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.129278] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1165.129596] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1165.131795] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02d7cb7-a533-4794-b9aa-6781eaac2dbd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.154246] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103699, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.197319] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dcb80c-2b50-42d2-a3e3-8ec559cb5b2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.200781] env[65107]: DEBUG nova.compute.manager [req-e62d44b4-e47e-4755-bb34-779ea52f7190 req-0f992fe9-0baa-4af7-93a3-701776a2d3fa service nova] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Detach interface failed, port_id=b633cd1f-f333-4c16-8e85-b43212472666, reason: Instance d52a955c-bc3c-41f2-b3ab-db9da613279e could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1165.231627] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1/volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.232163] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103700, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.232502] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31f993f2-3829-4239-afb5-60edbfd34df5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.253418] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1165.253418] env[65107]: value = "task-5103702" [ 1165.253418] env[65107]: _type = "Task" [ 1165.253418] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.267277] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103702, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.332664] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103701, 'name': ReconfigVM_Task, 'duration_secs': 0.380849} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.333027] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 7e72891b-4c20-4889-80c9-8bcb15373a2c/7e72891b-4c20-4889-80c9-8bcb15373a2c.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.333868] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3b6601f-e0c7-477e-ae3b-61f2ed111cb0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.341772] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1165.341772] env[65107]: value = "task-5103703" [ 1165.341772] env[65107]: _type = "Task" [ 1165.341772] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.351082] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103703, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.520724] env[65107]: DEBUG nova.network.neutron [-] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1165.554229] env[65107]: INFO nova.compute.manager [-] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Took 1.28 seconds to deallocate network for instance. [ 1165.570295] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d5766a03-054f-40ea-a57e-e640664ca683] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1165.598360] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fa042de-37a2-4158-92e4-9ae9bceb5bdc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.616354] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c2aa6f-6ff0-4d52-92c3-c5646f98df3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.642718] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103699, 'name': Rename_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.662900] env[65107]: DEBUG nova.compute.manager [req-741a4baa-3c0e-46de-9ed2-1a92c7651437 req-e3ff3f7e-18cb-46e7-96fb-362fad13fd67 service nova] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Detach interface failed, port_id=d04b7a1c-d3b0-4211-b44b-2e39791f7497, reason: Instance 0599df85-ad33-4bd5-b234-611db1ef73f0 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1165.682230] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103700, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.768787] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.853201] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103703, 'name': Rename_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.024572] env[65107]: INFO nova.compute.manager [-] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Took 1.92 seconds to deallocate network for instance. [ 1166.062237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.062237] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.062237] env[65107]: DEBUG nova.objects.instance [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'resources' on Instance uuid d52a955c-bc3c-41f2-b3ab-db9da613279e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.073098] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 23c3b100-30ac-44f9-8e2d-d8e3f2f4e151] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1166.142692] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103699, 'name': Rename_Task, 'duration_secs': 1.166646} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.142930] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1166.143204] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75e2b36f-b068-415d-8033-1bb2d7412c45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.153422] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1166.153422] env[65107]: value = "task-5103704" [ 1166.153422] env[65107]: _type = "Task" [ 1166.153422] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.164060] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.185372] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103700, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.267840] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103702, 'name': ReconfigVM_Task, 'duration_secs': 0.864953} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.269024] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1/volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1166.278218] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57445729-98f0-4999-b474-2ca14fa72648 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.298060] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1166.298060] env[65107]: value = "task-5103705" [ 1166.298060] env[65107]: _type = "Task" [ 1166.298060] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.308546] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103705, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.353096] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103703, 'name': Rename_Task, 'duration_secs': 0.958819} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.353532] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1166.353769] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f9a2469-ee65-4ad8-b391-b3601a041ea1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.361798] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1166.361798] env[65107]: value = "task-5103706" [ 1166.361798] env[65107]: _type = "Task" [ 1166.361798] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.379317] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.532180] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.578922] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 888d60d4-bc9c-48af-97f9-434883884e8a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1166.667913] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103704, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.687106] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103700, 'name': CloneVM_Task, 'duration_secs': 1.716625} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.690230] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Created linked-clone VM from snapshot [ 1166.691627] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc8e009-06de-4509-8045-88b8e6acc2e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.704028] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Uploading image e59d3649-e85b-4c3e-9b08-6bf703feaba9 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1166.723741] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1166.724317] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8ce61d3c-ecc2-4ae6-835f-cb5f1eb8389a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.733132] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1166.733132] env[65107]: value = "task-5103707" [ 1166.733132] env[65107]: _type = "Task" [ 1166.733132] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.749469] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103707, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.810751] env[65107]: DEBUG oslo_vmware.api [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103705, 'name': ReconfigVM_Task, 'duration_secs': 0.191203} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.814808] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1166.849273] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93148462-4b5d-4417-b6dd-a333c242c7bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.860348] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79128ec3-e34b-4a2a-a82d-20481624e3cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.877116] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103706, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.917519] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5773ed-1902-4dd9-984a-b252db5204cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.927914] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a375b0d-19c2-492d-8ab9-c3ed97402927 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.948668] env[65107]: DEBUG nova.compute.provider_tree [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.083097] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 70172f4f-2658-4bc1-857f-0dffde2d5ab0] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1167.165372] env[65107]: DEBUG oslo_vmware.api [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103704, 'name': PowerOnVM_Task, 'duration_secs': 0.548986} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.165703] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1167.165931] env[65107]: INFO nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Took 9.77 seconds to spawn the instance on the hypervisor. [ 1167.166131] env[65107]: DEBUG nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1167.167081] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0dc87ac-1bc8-4263-ba00-08b684e4cbc0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.245748] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103707, 'name': Destroy_Task} progress is 33%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.378376] env[65107]: DEBUG oslo_vmware.api [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103706, 'name': PowerOnVM_Task, 'duration_secs': 0.580835} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.378683] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1167.378969] env[65107]: INFO nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Took 7.60 seconds to spawn the instance on the hypervisor. [ 1167.379136] env[65107]: DEBUG nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1167.380072] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79e12f9-645d-48a5-bdda-bc2d27323300 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.451962] env[65107]: DEBUG nova.scheduler.client.report [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1167.586969] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: fead9fe9-23f4-4d13-bf44-be0727057ddd] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1167.688192] env[65107]: INFO nova.compute.manager [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Took 15.48 seconds to build instance. [ 1167.746084] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103707, 'name': Destroy_Task, 'duration_secs': 0.676987} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.746461] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Destroyed the VM [ 1167.746715] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1167.747075] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-34483e4d-ee16-4008-805f-16673eaa43bd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.755096] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1167.755096] env[65107]: value = "task-5103708" [ 1167.755096] env[65107]: _type = "Task" [ 1167.755096] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.764801] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103708, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.857597] env[65107]: DEBUG nova.objects.instance [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'flavor' on Instance uuid bef91b92-add1-4df9-bc24-dab15ce04338 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.904797] env[65107]: INFO nova.compute.manager [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Took 14.75 seconds to build instance. [ 1167.958784] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.897s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.961479] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.430s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.961788] env[65107]: DEBUG nova.objects.instance [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'resources' on Instance uuid 0599df85-ad33-4bd5-b234-611db1ef73f0 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.988458] env[65107]: INFO nova.scheduler.client.report [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance d52a955c-bc3c-41f2-b3ab-db9da613279e [ 1168.091332] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: afa1a4cc-aa43-4e10-982f-edaf5cdee584] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1168.123561] env[65107]: DEBUG nova.compute.manager [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Received event network-changed-13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1168.123880] env[65107]: DEBUG nova.compute.manager [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Refreshing instance network info cache due to event network-changed-13b14810-c694-4e47-b090-8c3bf39ad510. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1168.124160] env[65107]: DEBUG oslo_concurrency.lockutils [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Acquiring lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.124365] env[65107]: DEBUG oslo_concurrency.lockutils [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Acquired lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.124656] env[65107]: DEBUG nova.network.neutron [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Refreshing network info cache for port 13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1168.190715] env[65107]: DEBUG oslo_concurrency.lockutils [None req-269409d6-afe8-4537-8089-bd0f2b3510b5 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.991s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.270933] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103708, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.367335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01a97c40-600a-4e23-bab5-a60370d70673 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.924s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.407396] env[65107]: DEBUG oslo_concurrency.lockutils [None req-48ca426f-55e5-43bf-8a8d-152e5c28737b tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.270s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.496361] env[65107]: DEBUG oslo_concurrency.lockutils [None req-19001f15-ad4e-4f87-9992-fcb830d113b1 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "d52a955c-bc3c-41f2-b3ab-db9da613279e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.929s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.597437] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d2e7eee7-c9b2-49aa-92f9-1efd7c901509] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1168.632378] env[65107]: WARNING neutronclient.v2_0.client [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.634151] env[65107]: WARNING openstack [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.635298] env[65107]: WARNING openstack [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.647060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "7e72891b-4c20-4889-80c9-8bcb15373a2c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.647412] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.647697] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "7e72891b-4c20-4889-80c9-8bcb15373a2c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.651021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.651021] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.654969] env[65107]: INFO nova.compute.manager [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Terminating instance [ 1168.757599] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae92e480-37f2-4585-a251-b97f9498d06e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.778797] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd13ce8-88f2-4c8a-8baf-5634b4cad585 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.782181] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103708, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.813940] env[65107]: WARNING openstack [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.814340] env[65107]: WARNING openstack [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.821270] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1253136-afd5-4862-8964-fcfeb86e8971 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.830612] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3348958e-4279-41bd-864e-59b7fcf10566 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.848048] env[65107]: DEBUG nova.compute.provider_tree [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.869358] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.869604] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.869805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.869974] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.870587] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.872842] env[65107]: INFO nova.compute.manager [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Terminating instance [ 1168.897766] env[65107]: WARNING neutronclient.v2_0.client [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.898495] env[65107]: WARNING openstack [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.898840] env[65107]: WARNING openstack [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.982760] env[65107]: DEBUG nova.network.neutron [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updated VIF entry in instance network info cache for port 13b14810-c694-4e47-b090-8c3bf39ad510. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1168.983174] env[65107]: DEBUG nova.network.neutron [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [{"id": "13b14810-c694-4e47-b090-8c3bf39ad510", "address": "fa:16:3e:e7:23:0e", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13b14810-c6", "ovs_interfaceid": "13b14810-c694-4e47-b090-8c3bf39ad510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1169.105566] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3a975793-f480-4f54-85c9-2c2aea11c18a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1169.159351] env[65107]: DEBUG nova.compute.manager [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1169.159516] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1169.160519] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c7a83f-cb86-4fc1-bdb7-e44947473d6d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.169387] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.169651] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad559d19-cb63-4f82-82d2-f475732e4a19 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.177443] env[65107]: DEBUG oslo_vmware.api [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1169.177443] env[65107]: value = "task-5103709" [ 1169.177443] env[65107]: _type = "Task" [ 1169.177443] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.187628] env[65107]: DEBUG oslo_vmware.api [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.267670] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103708, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.355751] env[65107]: DEBUG nova.scheduler.client.report [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1169.359056] env[65107]: INFO nova.compute.manager [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Rebuilding instance [ 1169.376887] env[65107]: DEBUG nova.compute.manager [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1169.377181] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1169.378242] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a677860-e3bb-44df-87ab-ae3a806d13ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.389912] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.394421] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ceb19fce-5ec6-489e-9f0b-2b024bdb0bb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.403607] env[65107]: DEBUG oslo_vmware.api [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1169.403607] env[65107]: value = "task-5103710" [ 1169.403607] env[65107]: _type = "Task" [ 1169.403607] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.415300] env[65107]: DEBUG oslo_vmware.api [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103710, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.422408] env[65107]: DEBUG nova.compute.manager [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1169.423816] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ba66d9-15b2-454e-b56d-10b7aa1b8eb6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.486516] env[65107]: DEBUG oslo_concurrency.lockutils [req-313bf772-6eb8-44f4-987d-5b0fbfbf206d req-40ebfe96-754e-4fff-8e80-73c6187b255c service nova] Releasing lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1169.610677] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 4a42392a-a4b4-43b7-a410-0b69441435be] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1169.688069] env[65107]: DEBUG oslo_vmware.api [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103709, 'name': PowerOffVM_Task, 'duration_secs': 0.210314} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.688364] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1169.688639] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1169.688919] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc3e230f-9706-449c-84f3-80fbefd48ca9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.768274] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103708, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.862924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.901s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.885564] env[65107]: INFO nova.scheduler.client.report [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted allocations for instance 0599df85-ad33-4bd5-b234-611db1ef73f0 [ 1169.914937] env[65107]: DEBUG oslo_vmware.api [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103710, 'name': PowerOffVM_Task, 'duration_secs': 0.190962} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.915251] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1169.915488] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1169.915744] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1378d4fa-4955-4c25-b595-2261d1575fa2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.981084] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1169.981327] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1169.981580] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleting the datastore file [datastore2] 7e72891b-4c20-4889-80c9-8bcb15373a2c {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1169.981900] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9193b3d-8eac-4030-9d3d-fe26e779fcc5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.993335] env[65107]: DEBUG oslo_vmware.api [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1169.993335] env[65107]: value = "task-5103713" [ 1169.993335] env[65107]: _type = "Task" [ 1169.993335] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.999379] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1169.999742] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.000016] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore1] 1f3292b7-f6d7-43ca-9caf-8c55794e8311 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.000890] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c332afd7-1519-46f9-989f-f054971b0cf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.007463] env[65107]: DEBUG oslo_vmware.api [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.014447] env[65107]: DEBUG oslo_vmware.api [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1170.014447] env[65107]: value = "task-5103714" [ 1170.014447] env[65107]: _type = "Task" [ 1170.014447] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.025764] env[65107]: DEBUG oslo_vmware.api [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103714, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.114823] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 1a813da6-f777-4888-b2ae-fbad58e01f61] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1170.268756] env[65107]: DEBUG oslo_vmware.api [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103708, 'name': RemoveSnapshot_Task, 'duration_secs': 2.096024} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.268983] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1170.394401] env[65107]: DEBUG oslo_concurrency.lockutils [None req-cdb4c112-d7c0-44ec-87e4-7403a9f2dc10 tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "0599df85-ad33-4bd5-b234-611db1ef73f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.130s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.439323] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1170.439643] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7050d939-b3f4-4b43-9e33-11bfd4e52ca2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.448822] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1170.448822] env[65107]: value = "task-5103715" [ 1170.448822] env[65107]: _type = "Task" [ 1170.448822] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.458083] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103715, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.505966] env[65107]: DEBUG oslo_vmware.api [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204313} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.505966] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.506295] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1170.506399] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1170.506593] env[65107]: INFO nova.compute.manager [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Took 1.35 seconds to destroy the instance on the hypervisor. [ 1170.506919] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1170.507295] env[65107]: DEBUG nova.compute.manager [-] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1170.507295] env[65107]: DEBUG nova.network.neutron [-] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1170.507502] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.508059] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.508332] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.524547] env[65107]: DEBUG oslo_vmware.api [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103714, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18553} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.524780] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.524970] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1170.525170] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1170.525358] env[65107]: INFO nova.compute.manager [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1170.525658] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1170.525874] env[65107]: DEBUG nova.compute.manager [-] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1170.525972] env[65107]: DEBUG nova.network.neutron [-] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1170.526241] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.526781] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.527144] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.587047] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.595974] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.618790] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: ac631507-8abf-4019-bdd5-7684dd51d1e9] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1170.774114] env[65107]: WARNING nova.compute.manager [None req-c283bf51-6502-4b7f-9c5e-7fd501da5845 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Image not found during snapshot: nova.exception.ImageNotFound: Image e59d3649-e85b-4c3e-9b08-6bf703feaba9 could not be found. [ 1170.858626] env[65107]: DEBUG nova.compute.manager [req-53673013-6be3-4f76-80f8-6b52e3d7b6c7 req-f97da0a8-9ec3-44a2-af8f-713ec0381ad6 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Received event network-vif-deleted-ebe5fc14-2b6a-40f0-80c2-2f843e929082 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1170.858877] env[65107]: INFO nova.compute.manager [req-53673013-6be3-4f76-80f8-6b52e3d7b6c7 req-f97da0a8-9ec3-44a2-af8f-713ec0381ad6 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Neutron deleted interface ebe5fc14-2b6a-40f0-80c2-2f843e929082; detaching it from the instance and deleting it from the info cache [ 1170.859114] env[65107]: DEBUG nova.network.neutron [req-53673013-6be3-4f76-80f8-6b52e3d7b6c7 req-f97da0a8-9ec3-44a2-af8f-713ec0381ad6 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.882679] env[65107]: DEBUG nova.compute.manager [req-304c0346-008f-4c20-a512-ef494e8f7cda req-2cc15acf-3169-4850-b10f-faefd5880162 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Received event network-vif-deleted-c0845e53-b126-43de-a3da-5356c9c0a9c7 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1170.882958] env[65107]: INFO nova.compute.manager [req-304c0346-008f-4c20-a512-ef494e8f7cda req-2cc15acf-3169-4850-b10f-faefd5880162 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Neutron deleted interface c0845e53-b126-43de-a3da-5356c9c0a9c7; detaching it from the instance and deleting it from the info cache [ 1170.883216] env[65107]: DEBUG nova.network.neutron [req-304c0346-008f-4c20-a512-ef494e8f7cda req-2cc15acf-3169-4850-b10f-faefd5880162 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.960235] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103715, 'name': PowerOffVM_Task, 'duration_secs': 0.425242} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.960549] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1171.018780] env[65107]: INFO nova.compute.manager [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Detaching volume f451fb1b-2805-4784-8cf8-a2cce68d18d1 [ 1171.054660] env[65107]: INFO nova.virt.block_device [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Attempting to driver detach volume f451fb1b-2805-4784-8cf8-a2cce68d18d1 from mountpoint /dev/sdb [ 1171.054930] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1171.055132] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1171.056090] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277f804b-ae31-4695-b622-5621306801f0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.083566] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499215b4-7606-4865-8b68-b6f6e692fd00 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.093717] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9424be1c-8d4d-433a-9afc-ac1a71a449c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.116950] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3a4dd0-1fd1-4423-b7a4-386b15a71798 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.121882] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: cddb14db-34c3-4783-b21e-e0b4623a3a9e] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1171.138392] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The volume has not been displaced from its original location: [datastore2] volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1/volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1171.143856] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1171.144251] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efd54163-961c-443e-a628-bf0baa401f94 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.165543] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1171.165543] env[65107]: value = "task-5103716" [ 1171.165543] env[65107]: _type = "Task" [ 1171.165543] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.174512] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103716, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.259746] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.260190] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.260571] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.260850] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.261097] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.263624] env[65107]: INFO nova.compute.manager [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Terminating instance [ 1171.314935] env[65107]: DEBUG nova.network.neutron [-] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1171.333835] env[65107]: DEBUG nova.network.neutron [-] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1171.362650] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94ce3953-a18a-46f4-8a8f-717d895ca1b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.374094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afea277-e86d-4465-bc6f-b5873a07150e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.386284] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-853df9ef-cb11-486b-b2f6-bf1687eb00af {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.395857] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9568b6b4-3b1c-45da-af2e-eb99a870d7a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.418884] env[65107]: DEBUG nova.compute.manager [req-53673013-6be3-4f76-80f8-6b52e3d7b6c7 req-f97da0a8-9ec3-44a2-af8f-713ec0381ad6 service nova] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Detach interface failed, port_id=ebe5fc14-2b6a-40f0-80c2-2f843e929082, reason: Instance 7e72891b-4c20-4889-80c9-8bcb15373a2c could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1171.444596] env[65107]: DEBUG nova.compute.manager [req-304c0346-008f-4c20-a512-ef494e8f7cda req-2cc15acf-3169-4850-b10f-faefd5880162 service nova] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Detach interface failed, port_id=c0845e53-b126-43de-a3da-5356c9c0a9c7, reason: Instance 1f3292b7-f6d7-43ca-9caf-8c55794e8311 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1171.583645] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.584340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.584632] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.584920] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.585124] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.587616] env[65107]: INFO nova.compute.manager [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Terminating instance [ 1171.624951] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: fb4229c6-b80d-49e9-a68b-baa12c0004ed] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1171.677970] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103716, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.768047] env[65107]: DEBUG nova.compute.manager [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1171.768299] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1171.769283] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144bd5e9-6dcb-46cd-a60d-52a8c8430d08 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.779291] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1171.779652] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52a04cc6-f692-45a6-9424-c1fff4dc2f17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.786955] env[65107]: DEBUG oslo_vmware.api [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1171.786955] env[65107]: value = "task-5103717" [ 1171.786955] env[65107]: _type = "Task" [ 1171.786955] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.798851] env[65107]: DEBUG oslo_vmware.api [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.818792] env[65107]: INFO nova.compute.manager [-] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Took 1.31 seconds to deallocate network for instance. [ 1171.836787] env[65107]: INFO nova.compute.manager [-] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Took 1.31 seconds to deallocate network for instance. [ 1172.091498] env[65107]: DEBUG nova.compute.manager [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1172.091740] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1172.092709] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca65b1a2-6bb5-4b65-967c-3b8a1f41fa5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.100889] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.101202] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fcb2b4f-6159-4241-b7ac-54a7c0c38074 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.108295] env[65107]: DEBUG oslo_vmware.api [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1172.108295] env[65107]: value = "task-5103718" [ 1172.108295] env[65107]: _type = "Task" [ 1172.108295] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.118140] env[65107]: DEBUG oslo_vmware.api [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.127989] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 9fa949ea-358c-46e7-a0f2-4c3275493b64] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1172.179039] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103716, 'name': ReconfigVM_Task, 'duration_secs': 0.570833} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.179247] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1172.184696] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb530678-bb4f-481f-90d6-2e0216db5c89 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.203860] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1172.203860] env[65107]: value = "task-5103719" [ 1172.203860] env[65107]: _type = "Task" [ 1172.203860] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.217301] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103719, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.298783] env[65107]: DEBUG oslo_vmware.api [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103717, 'name': PowerOffVM_Task, 'duration_secs': 0.21569} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.299026] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.300247] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1172.300247] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41b49f1b-6d68-45ee-8ae8-8ab1d6b02cbf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.326700] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.327054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.327282] env[65107]: DEBUG nova.objects.instance [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'resources' on Instance uuid 7e72891b-4c20-4889-80c9-8bcb15373a2c {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.342755] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.388529] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1172.388915] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1172.389145] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleting the datastore file [datastore1] 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.389460] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47cb0301-d211-41ca-9cdb-6933d6504836 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.399025] env[65107]: DEBUG oslo_vmware.api [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for the task: (returnval){ [ 1172.399025] env[65107]: value = "task-5103721" [ 1172.399025] env[65107]: _type = "Task" [ 1172.399025] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.408774] env[65107]: DEBUG oslo_vmware.api [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.619580] env[65107]: DEBUG oslo_vmware.api [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103718, 'name': PowerOffVM_Task, 'duration_secs': 0.206575} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.619906] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.620014] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1172.620274] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d73333c-764c-44bc-9889-01fad880b62e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.631622] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3f34d6a2-0370-4239-9556-b4333e454d58] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1172.687362] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1172.687701] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1172.687918] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleting the datastore file [datastore2] 4f93a52c-9c40-47e5-a34f-b19aea34d7a9 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.688266] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae022b12-40a1-4920-8f35-2f3a820d7787 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.696330] env[65107]: DEBUG oslo_vmware.api [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1172.696330] env[65107]: value = "task-5103723" [ 1172.696330] env[65107]: _type = "Task" [ 1172.696330] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.705323] env[65107]: DEBUG oslo_vmware.api [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.714484] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103719, 'name': ReconfigVM_Task, 'duration_secs': 0.158367} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.714612] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1172.909746] env[65107]: DEBUG oslo_vmware.api [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Task: {'id': task-5103721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115875} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.910011] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1172.910232] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1172.910427] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1172.910602] env[65107]: INFO nova.compute.manager [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1172.910853] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1172.911056] env[65107]: DEBUG nova.compute.manager [-] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1172.911157] env[65107]: DEBUG nova.network.neutron [-] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1172.911401] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1172.911965] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.912280] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.949923] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.128239] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42be0ce8-6f12-4ee8-87b4-8674c2c756a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.135233] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 40e17da0-b459-4619-8e9e-18c3093328c4] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1173.138249] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce21954-5dab-42df-a9af-e9a6b8d36109 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.174947] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52d454b-f157-4755-aada-b8f8cc07630a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.186553] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273a1a96-607b-43b6-a0d4-a6d402f11207 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.214612] env[65107]: DEBUG nova.compute.provider_tree [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.222350] env[65107]: DEBUG nova.compute.manager [req-c16aa7b2-0d80-45c4-ac2a-d741d3db669c req-80b6d073-3363-42b9-9936-a4c2b40817c6 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Received event network-vif-deleted-47322e57-eaa6-4bb8-98ec-3098749bd52c {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1173.222628] env[65107]: INFO nova.compute.manager [req-c16aa7b2-0d80-45c4-ac2a-d741d3db669c req-80b6d073-3363-42b9-9936-a4c2b40817c6 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Neutron deleted interface 47322e57-eaa6-4bb8-98ec-3098749bd52c; detaching it from the instance and deleting it from the info cache [ 1173.222819] env[65107]: DEBUG nova.network.neutron [req-c16aa7b2-0d80-45c4-ac2a-d741d3db669c req-80b6d073-3363-42b9-9936-a4c2b40817c6 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1173.233241] env[65107]: DEBUG oslo_vmware.api [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142384} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.233626] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.233793] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.233967] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.234175] env[65107]: INFO nova.compute.manager [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1173.234496] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1173.236238] env[65107]: DEBUG nova.compute.manager [-] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1173.236238] env[65107]: DEBUG nova.network.neutron [-] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1173.236238] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.236238] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.240257] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.284337] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.643149] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 64b4fe7b-3768-4fd8-a6bb-5c8f96aff24e] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1173.693469] env[65107]: DEBUG nova.network.neutron [-] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1173.726025] env[65107]: DEBUG nova.scheduler.client.report [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1173.730075] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df64d165-9332-482a-b153-406fd1b620e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.740131] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e4731c-8821-4281-a9a5-9a4a6651a789 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.777725] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.778249] env[65107]: DEBUG nova.compute.manager [req-c16aa7b2-0d80-45c4-ac2a-d741d3db669c req-80b6d073-3363-42b9-9936-a4c2b40817c6 service nova] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Detach interface failed, port_id=47322e57-eaa6-4bb8-98ec-3098749bd52c, reason: Instance 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1173.778681] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87fcd989-8372-41a9-af7d-c3522e379dc2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.787052] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1173.787052] env[65107]: value = "task-5103724" [ 1173.787052] env[65107]: _type = "Task" [ 1173.787052] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.795770] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.033448] env[65107]: DEBUG nova.network.neutron [-] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1174.146111] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: ee50d08a-57fc-4c05-96fe-a11fe2708165] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1174.196568] env[65107]: INFO nova.compute.manager [-] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Took 1.29 seconds to deallocate network for instance. [ 1174.232184] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.234903] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.892s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.235859] env[65107]: DEBUG nova.objects.instance [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'resources' on Instance uuid 1f3292b7-f6d7-43ca-9caf-8c55794e8311 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.263378] env[65107]: INFO nova.scheduler.client.report [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted allocations for instance 7e72891b-4c20-4889-80c9-8bcb15373a2c [ 1174.298069] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1174.298279] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1174.298730] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1174.299447] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502940c9-5c5d-4656-a7f6-5eb77888739c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.320102] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c815c6-ed04-400b-b15f-5f2b2e9354c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.330204] env[65107]: WARNING nova.virt.vmwareapi.driver [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1174.330516] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.331501] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad717355-ac8f-48cb-b423-7ee06fd7fe6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.340532] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.340728] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6150603a-e204-4175-a938-3b0db72f6d45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.427369] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.427796] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.427998] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore1] bef91b92-add1-4df9-bc24-dab15ce04338 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.428312] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65c2fbbc-3194-4958-962b-8650bffac503 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.436126] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1174.436126] env[65107]: value = "task-5103726" [ 1174.436126] env[65107]: _type = "Task" [ 1174.436126] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.446197] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.536520] env[65107]: INFO nova.compute.manager [-] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Took 1.30 seconds to deallocate network for instance. [ 1174.649848] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 88cac920-3c1e-41dc-9e52-f51f48a80f56] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1174.703203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.773547] env[65107]: DEBUG oslo_concurrency.lockutils [None req-afbdc8fa-1fb2-4ec5-83d0-cc7796eacc96 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "7e72891b-4c20-4889-80c9-8bcb15373a2c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.126s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.929599] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4f07e6-5ac3-4f2d-ab03-e2206ead030d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.940735] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d521c5-ef22-489c-92e2-e17584ef2399 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.949450] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147147} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.977278] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1174.977499] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1174.977674] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1174.980968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d62c5c-6ca9-42ff-824e-3f8427474925 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.989615] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd1ce9f-b487-4eb1-84a1-9e9db453a211 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.004866] env[65107]: DEBUG nova.compute.provider_tree [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.049444] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.153643] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 9a46ff51-4035-4709-ae6c-0fb5397200ab] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1175.245801] env[65107]: DEBUG nova.compute.manager [req-f2a0790e-88e8-4952-af21-068685c77712 req-45ce8d0e-148f-494a-865e-4de1ef3c7d6c service nova] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Received event network-vif-deleted-079f6503-3319-4d8a-b94b-843464a762f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1175.486200] env[65107]: INFO nova.virt.block_device [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Booting with volume f451fb1b-2805-4784-8cf8-a2cce68d18d1 at /dev/sdb [ 1175.508414] env[65107]: DEBUG nova.scheduler.client.report [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.519780] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c219f2b-cbc8-4d2e-b6a6-2ae6f55943e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.531655] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682bb32a-aabe-4f14-9809-8395e2290c71 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.571223] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ceee387-5e71-42a9-a9af-06efcbda32d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.582164] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5b678b-4acd-4fc1-bb77-d20d59df42f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.618254] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbcb082-a9fc-46ca-bf6e-12bf8880d7e3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.625791] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1253776-1dca-46b1-b8b4-4d09753cbe6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.639704] env[65107]: DEBUG nova.virt.block_device [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating existing volume attachment record: 443e208b-eb5f-47fe-91fe-85eb84809692 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1175.656832] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 21c471cd-2435-4e4e-82cc-8cdf03f1fa6c] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1176.014859] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.017874] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.315s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.018544] env[65107]: DEBUG nova.objects.instance [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lazy-loading 'resources' on Instance uuid 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.037770] env[65107]: INFO nova.scheduler.client.report [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance 1f3292b7-f6d7-43ca-9caf-8c55794e8311 [ 1176.160615] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 40474981-eeef-492e-8a8b-aaea7f554c02] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1176.547145] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ed14f355-c4c3-4ded-a6be-95bc118949cd tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "1f3292b7-f6d7-43ca-9caf-8c55794e8311" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.677s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.664578] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 2d0a93a4-59a8-4788-93ce-c34af1432a4f] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1176.710343] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d1f98f-cdee-400f-a994-9cb40226a174 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.721105] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fa7eb5-760e-452e-afcd-5d36769eb3b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.760845] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f643f9-073b-410d-9411-0a4e36683fa2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.769819] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fd66a8-4454-410b-b672-1cbf434612b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.785817] env[65107]: DEBUG nova.compute.provider_tree [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.168158] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 805e2210-fb58-48da-be63-5d548b2f8a2f] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1177.210634] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.210946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.289532] env[65107]: DEBUG nova.scheduler.client.report [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1177.671638] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 5fccc802-f5ac-4450-8246-4cf9a5371046] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1177.714325] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1177.765239] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1177.765569] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1177.765740] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1177.765921] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1177.766075] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1177.766225] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1177.766430] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1177.766589] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1177.766752] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1177.767039] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1177.767251] env[65107]: DEBUG nova.virt.hardware [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1177.768419] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a85b39-0a39-469c-890e-c4109061c718 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.776774] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e857ed4d-7034-4c35-a9ee-856b68727e76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.793898] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:32:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40c68565-653b-4c27-83db-52fff0de54f0', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.801259] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1177.801999] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.803965] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1177.804471] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.755s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.804704] env[65107]: DEBUG nova.objects.instance [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lazy-loading 'resources' on Instance uuid 4f93a52c-9c40-47e5-a34f-b19aea34d7a9 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.805656] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae117336-0a07-4eff-83af-cad7df7a9dd0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.824042] env[65107]: INFO nova.scheduler.client.report [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Deleted allocations for instance 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d [ 1177.830804] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.830804] env[65107]: value = "task-5103727" [ 1177.830804] env[65107]: _type = "Task" [ 1177.830804] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.840072] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103727, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.993586] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2577246-7f17-4633-9582-9aefd7f3b655 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.002156] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84007dc8-bb72-4bd3-9dea-5903962d3d2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.032772] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031a1926-e982-42e7-b393-cf4147955646 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.041525] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f429a5-ff8e-497c-b39f-f5f546a14a3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.056863] env[65107]: DEBUG nova.compute.provider_tree [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.175233] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 780f910f-4f73-41ba-a795-0daae3097314] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1178.240513] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.331806] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c07d18e5-f9ca-4e55-b98c-2857ea9e532a tempest-AttachInterfacesTestJSON-1317707612 tempest-AttachInterfacesTestJSON-1317707612-project-member] Lock "709eaf30-f1a6-49c2-a0fd-a47eee8aef2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.072s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.344162] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103727, 'name': CreateVM_Task, 'duration_secs': 0.347577} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.344337] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1178.344935] env[65107]: WARNING neutronclient.v2_0.client [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1178.345327] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.345485] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.345830] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1178.346167] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e5c7ec9-8840-4410-81bc-caadeb288444 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.353555] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1178.353555] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cd703-97fa-152e-7f17-ec064be2d3d4" [ 1178.353555] env[65107]: _type = "Task" [ 1178.353555] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.364514] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cd703-97fa-152e-7f17-ec064be2d3d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.560469] env[65107]: DEBUG nova.scheduler.client.report [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1178.679031] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 179e613b-e5a2-4fbc-8fa6-b72769425ff3] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1178.866599] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]523cd703-97fa-152e-7f17-ec064be2d3d4, 'name': SearchDatastore_Task, 'duration_secs': 0.012105} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.866943] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.867212] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.867456] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.867609] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.867788] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.868094] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7afdc610-00e8-4e18-aeb7-ae14516acb0b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.879562] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.880473] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.880892] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f283edef-f082-443f-beed-8770f6ffb54c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.888957] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1178.888957] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520b3b0c-cb9b-6cf9-a82e-5fbc665ed35a" [ 1178.888957] env[65107]: _type = "Task" [ 1178.888957] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.898078] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520b3b0c-cb9b-6cf9-a82e-5fbc665ed35a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.066084] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.261s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.068927] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.828s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1179.071133] env[65107]: INFO nova.compute.claims [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1179.102675] env[65107]: INFO nova.scheduler.client.report [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleted allocations for instance 4f93a52c-9c40-47e5-a34f-b19aea34d7a9 [ 1179.182081] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: a4f43fbc-63e3-4b69-bdb1-7aaf47aae090] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1179.400901] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520b3b0c-cb9b-6cf9-a82e-5fbc665ed35a, 'name': SearchDatastore_Task, 'duration_secs': 0.011322} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.401734] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-957c1f38-4322-4889-a294-41b988818268 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.407939] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1179.407939] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e03e9d-8fb9-69cc-892b-93690398ccc0" [ 1179.407939] env[65107]: _type = "Task" [ 1179.407939] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.417842] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e03e9d-8fb9-69cc-892b-93690398ccc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.611558] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a45c9879-a98d-4074-85b9-b2d751132ce6 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "4f93a52c-9c40-47e5-a34f-b19aea34d7a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.027s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.687697] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 82e8c093-b718-4d38-9682-ba2e710e4b93] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1179.922785] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e03e9d-8fb9-69cc-892b-93690398ccc0, 'name': SearchDatastore_Task, 'duration_secs': 0.012663} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.923135] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1179.923425] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1179.923701] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8259a7a-2d6c-4e4a-ad9e-dc0ed8465b33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.932145] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1179.932145] env[65107]: value = "task-5103728" [ 1179.932145] env[65107]: _type = "Task" [ 1179.932145] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.943173] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103728, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.191416] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: bfe544ec-9b08-4118-a940-a51520ecaac0] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1180.311586] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4885a137-79fb-4caf-b34e-eb38f3aff6c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.326913] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b348a18-d138-40ef-935e-d135dbdd2c3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.388351] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acbf962-488d-4ab3-b671-c6a0da23f3ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.403968] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f40beb-acc0-4adb-9b3a-87f2a5b386df {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.424931] env[65107]: DEBUG nova.compute.provider_tree [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.447792] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103728, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.699138] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d2d0e18c-4839-416c-b363-b93996e8ba7f] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1180.928782] env[65107]: DEBUG nova.scheduler.client.report [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.949639] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103728, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635294} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.950881] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1180.951153] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1180.951445] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73630897-0e1c-4a56-b03f-b7cfeeb423bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.961983] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1180.961983] env[65107]: value = "task-5103729" [ 1180.961983] env[65107]: _type = "Task" [ 1180.961983] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.975255] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103729, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.194059] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.194716] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.201776] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d65e98bb-f54d-4803-95ca-0fc9b6c184bc] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1181.440973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.441654] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1181.473235] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103729, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072444} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.474047] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1181.475141] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3221e6d-0e89-42ec-8734-6edbae5b3cb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.500161] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.501311] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5249623c-5ecb-4a62-8491-ab5a708cd441 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.524780] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1181.524780] env[65107]: value = "task-5103730" [ 1181.524780] env[65107]: _type = "Task" [ 1181.524780] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.535785] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103730, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.697067] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1181.705611] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 369caf24-48a6-4640-8601-e30fd47453da] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1181.947560] env[65107]: DEBUG nova.compute.utils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1181.949112] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1181.949379] env[65107]: DEBUG nova.network.neutron [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1181.949840] env[65107]: WARNING neutronclient.v2_0.client [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.950359] env[65107]: WARNING neutronclient.v2_0.client [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.951235] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.951774] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.963434] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1182.001560] env[65107]: DEBUG nova.policy [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1182.035042] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103730, 'name': ReconfigVM_Task, 'duration_secs': 0.511422} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.035412] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfigured VM instance instance-0000006c to attach disk [datastore2] bef91b92-add1-4df9-bc24-dab15ce04338/bef91b92-add1-4df9-bc24-dab15ce04338.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.036512] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'boot_index': 0, 'encrypted': False, 'device_name': '/dev/sda', 'size': 0, 'encryption_secret_uuid': None, 'encryption_options': None, 'encryption_format': None, 'guest_format': None, 'device_type': 'disk', 'image_id': 'e439aeac-2298-462c-be63-8218195135cf'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'}, 'disk_bus': None, 'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'attachment_id': '443e208b-eb5f-47fe-91fe-85eb84809692', 'mount_device': '/dev/sdb', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=65107) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1182.036738] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1182.036949] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1182.037860] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85661725-b814-4379-9ce4-7d5e63684ef5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.058054] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50138913-fb4e-4863-af99-2ffa6905ec29 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.085566] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1/volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1182.086074] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe7c4cbf-c74d-4765-adaf-6aac60cd568b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.105759] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1182.105759] env[65107]: value = "task-5103731" [ 1182.105759] env[65107]: _type = "Task" [ 1182.105759] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.114998] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103731, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.209736] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: ef417f32-8c0e-4a62-9078-cb2bf95b85f8] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1182.226471] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.226791] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.228386] env[65107]: INFO nova.compute.claims [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1182.343763] env[65107]: DEBUG nova.network.neutron [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Successfully created port: c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1182.616988] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103731, 'name': ReconfigVM_Task, 'duration_secs': 0.359815} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.617333] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1/volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.622329] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-013084e1-86af-432b-9173-c963976fe17c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.638829] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1182.638829] env[65107]: value = "task-5103732" [ 1182.638829] env[65107]: _type = "Task" [ 1182.638829] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.647820] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103732, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.713935] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 71e6e280-50c3-4655-a6a3-40ea4c650d96] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1182.973029] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1183.001545] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1183.001912] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1183.002118] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1183.002345] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1183.002513] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1183.002663] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1183.002878] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1183.003087] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1183.003267] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1183.003431] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1183.003605] env[65107]: DEBUG nova.virt.hardware [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1183.004588] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c585714-2012-4acf-8110-1ee3233c60ee {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.013622] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26d7d06-ce4c-4143-931e-8546b7fffc80 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.149155] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103732, 'name': ReconfigVM_Task, 'duration_secs': 0.151482} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.149497] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1183.150129] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7516cfca-7814-4a83-ac35-518232fe0443 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.159473] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1183.159473] env[65107]: value = "task-5103733" [ 1183.159473] env[65107]: _type = "Task" [ 1183.159473] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.171425] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103733, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.217576] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 7062b119-01a4-45f4-bb52-212178382271] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1183.427667] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f03786-e2ba-4f31-abc3-ac6d11caa63d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.436111] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48eb264-d0d1-46ad-80f9-834f1ee4c53b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.469299] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b6b782-2451-40a9-850c-745f41077ad9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.478126] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ffde68-be6d-407f-94c8-d9c31834d29e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.493516] env[65107]: DEBUG nova.compute.provider_tree [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.671029] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103733, 'name': Rename_Task, 'duration_secs': 0.162965} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.671361] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1183.671638] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1cfceab-2105-49e2-b1df-8e43c8357533 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.680147] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1183.680147] env[65107]: value = "task-5103734" [ 1183.680147] env[65107]: _type = "Task" [ 1183.680147] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.690112] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.720910] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 8306ff90-b8bd-4270-8133-96abe483156b] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1183.764862] env[65107]: DEBUG nova.compute.manager [req-7225dee6-0475-4327-856f-9e2c0307a8fc req-c5bb9c12-9f53-4326-8920-428ba67948cd service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Received event network-vif-plugged-c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1183.765038] env[65107]: DEBUG oslo_concurrency.lockutils [req-7225dee6-0475-4327-856f-9e2c0307a8fc req-c5bb9c12-9f53-4326-8920-428ba67948cd service nova] Acquiring lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.765177] env[65107]: DEBUG oslo_concurrency.lockutils [req-7225dee6-0475-4327-856f-9e2c0307a8fc req-c5bb9c12-9f53-4326-8920-428ba67948cd service nova] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.765498] env[65107]: DEBUG oslo_concurrency.lockutils [req-7225dee6-0475-4327-856f-9e2c0307a8fc req-c5bb9c12-9f53-4326-8920-428ba67948cd service nova] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.765567] env[65107]: DEBUG nova.compute.manager [req-7225dee6-0475-4327-856f-9e2c0307a8fc req-c5bb9c12-9f53-4326-8920-428ba67948cd service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] No waiting events found dispatching network-vif-plugged-c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1183.765976] env[65107]: WARNING nova.compute.manager [req-7225dee6-0475-4327-856f-9e2c0307a8fc req-c5bb9c12-9f53-4326-8920-428ba67948cd service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Received unexpected event network-vif-plugged-c54e07bb-9d9b-4424-84ec-301f0abb7e4e for instance with vm_state building and task_state spawning. [ 1183.872232] env[65107]: DEBUG nova.network.neutron [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Successfully updated port: c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1183.996682] env[65107]: DEBUG nova.scheduler.client.report [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1184.192257] env[65107]: DEBUG oslo_vmware.api [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103734, 'name': PowerOnVM_Task, 'duration_secs': 0.462708} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.192584] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1184.192922] env[65107]: DEBUG nova.compute.manager [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1184.194923] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0b2008-3892-4aa3-a70e-ef3bb09c4d3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.224669] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: fdb5603d-24b0-4fdd-b8cc-25c0ebee5311] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1184.375806] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-4a6894d7-233a-4c5c-8ded-5b80b9beadc7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.375950] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-4a6894d7-233a-4c5c-8ded-5b80b9beadc7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1184.376156] env[65107]: DEBUG nova.network.neutron [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1184.502931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.503469] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1184.715880] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.716173] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.716355] env[65107]: DEBUG nova.objects.instance [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65107) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1184.728263] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: b779b5ad-7d93-47c2-b824-6d76246c00f5] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1184.879011] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.879420] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.916653] env[65107]: DEBUG nova.network.neutron [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1184.936080] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.936483] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.984247] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.984615] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.984810] env[65107]: INFO nova.compute.manager [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Shelving [ 1185.003222] env[65107]: WARNING neutronclient.v2_0.client [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1185.003883] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.004244] env[65107]: WARNING openstack [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.013515] env[65107]: DEBUG nova.compute.utils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1185.015457] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1185.015457] env[65107]: DEBUG nova.network.neutron [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1185.015645] env[65107]: WARNING neutronclient.v2_0.client [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1185.015945] env[65107]: WARNING neutronclient.v2_0.client [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1185.016916] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1185.016916] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1185.069463] env[65107]: DEBUG nova.policy [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b73837038b2c4e678c0a7112f6a61f38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd28d5a8ab6f747628edf3aee83355feb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1185.098107] env[65107]: DEBUG nova.network.neutron [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Updating instance_info_cache with network_info: [{"id": "c54e07bb-9d9b-4424-84ec-301f0abb7e4e", "address": "fa:16:3e:79:98:02", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc54e07bb-9d", "ovs_interfaceid": "c54e07bb-9d9b-4424-84ec-301f0abb7e4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1185.231181] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: e3510c31-d6be-4e9f-a0a2-a662123861e9] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1185.395042] env[65107]: DEBUG nova.network.neutron [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Successfully created port: 3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1185.524770] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1185.600618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-4a6894d7-233a-4c5c-8ded-5b80b9beadc7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1185.600998] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Instance network_info: |[{"id": "c54e07bb-9d9b-4424-84ec-301f0abb7e4e", "address": "fa:16:3e:79:98:02", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc54e07bb-9d", "ovs_interfaceid": "c54e07bb-9d9b-4424-84ec-301f0abb7e4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1185.601494] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:98:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c54e07bb-9d9b-4424-84ec-301f0abb7e4e', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1185.609198] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1185.609427] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1185.609694] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60d68147-f5c1-4219-9b8f-e97a5b84045b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.631986] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1185.631986] env[65107]: value = "task-5103735" [ 1185.631986] env[65107]: _type = "Task" [ 1185.631986] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.640854] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103735, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.729326] env[65107]: DEBUG oslo_concurrency.lockutils [None req-03aa5abe-fbae-4aa0-901c-701bbc039329 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1185.734824] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.827431] env[65107]: DEBUG nova.compute.manager [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Received event network-changed-c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1185.827965] env[65107]: DEBUG nova.compute.manager [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Refreshing instance network info cache due to event network-changed-c54e07bb-9d9b-4424-84ec-301f0abb7e4e. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1185.828369] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Acquiring lock "refresh_cache-4a6894d7-233a-4c5c-8ded-5b80b9beadc7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.828599] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Acquired lock "refresh_cache-4a6894d7-233a-4c5c-8ded-5b80b9beadc7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.828888] env[65107]: DEBUG nova.network.neutron [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Refreshing network info cache for port c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1185.994852] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.995130] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ce55f01-15b1-446e-bf7c-e7cf26b6b7f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.005258] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1186.005258] env[65107]: value = "task-5103736" [ 1186.005258] env[65107]: _type = "Task" [ 1186.005258] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.016098] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.143118] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103735, 'name': CreateVM_Task, 'duration_secs': 0.35699} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.143360] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1186.143916] env[65107]: WARNING neutronclient.v2_0.client [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.144306] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.144507] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.144831] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1186.145116] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6cf5fdf-a109-49ac-b093-34ba71c8eeb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.150523] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1186.150523] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52db939c-2b49-154a-ea2c-3af0d3dfbca6" [ 1186.150523] env[65107]: _type = "Task" [ 1186.150523] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.161142] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52db939c-2b49-154a-ea2c-3af0d3dfbca6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.332779] env[65107]: WARNING neutronclient.v2_0.client [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.333695] env[65107]: WARNING openstack [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.334089] env[65107]: WARNING openstack [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.482614] env[65107]: WARNING openstack [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.483082] env[65107]: WARNING openstack [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.517549] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103736, 'name': PowerOffVM_Task, 'duration_secs': 0.225111} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.517549] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.517549] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c078dbf2-d2a9-4431-ac34-628f747e6995 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.542951] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1186.546183] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a331cb0-0873-4ecd-ab88-2400a25751bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.557815] env[65107]: WARNING neutronclient.v2_0.client [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.558512] env[65107]: WARNING openstack [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.558857] env[65107]: WARNING openstack [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.576276] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1186.576794] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1186.576794] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1186.576967] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1186.577103] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1186.577259] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1186.577510] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1186.577690] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1186.577885] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1186.578100] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1186.578314] env[65107]: DEBUG nova.virt.hardware [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1186.579303] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e842daa0-c597-4d1b-9de6-763a5923d55a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.588527] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee82714c-e5b8-4c4a-bb0b-84cc76a3bb1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.650430] env[65107]: DEBUG nova.network.neutron [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Updated VIF entry in instance network info cache for port c54e07bb-9d9b-4424-84ec-301f0abb7e4e. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1186.650888] env[65107]: DEBUG nova.network.neutron [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Updating instance_info_cache with network_info: [{"id": "c54e07bb-9d9b-4424-84ec-301f0abb7e4e", "address": "fa:16:3e:79:98:02", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc54e07bb-9d", "ovs_interfaceid": "c54e07bb-9d9b-4424-84ec-301f0abb7e4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1186.665120] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52db939c-2b49-154a-ea2c-3af0d3dfbca6, 'name': SearchDatastore_Task, 'duration_secs': 0.011386} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.666491] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1186.666875] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1186.667224] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.667474] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.667860] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1186.668745] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b6474f2-f20a-4163-a779-e2d585aa79a2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.681207] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1186.681575] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1186.682509] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4af284db-bf6f-42fa-85f3-2989065c6bf4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.691140] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1186.691140] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52600fcd-a9c7-69e2-a568-75c304acd7c4" [ 1186.691140] env[65107]: _type = "Task" [ 1186.691140] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.700222] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52600fcd-a9c7-69e2-a568-75c304acd7c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.922827] env[65107]: DEBUG nova.network.neutron [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Successfully updated port: 3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1187.068236] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1187.068673] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d022e9b6-b8e9-4ba1-b704-fb2a5ad4bae0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.077722] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1187.077722] env[65107]: value = "task-5103737" [ 1187.077722] env[65107]: _type = "Task" [ 1187.077722] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.087343] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103737, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.157380] env[65107]: DEBUG oslo_concurrency.lockutils [req-3a128960-0a31-46a4-92e3-b0b3059cc090 req-b86abb5d-092e-4446-b627-fde6b780c179 service nova] Releasing lock "refresh_cache-4a6894d7-233a-4c5c-8ded-5b80b9beadc7" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.202863] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52600fcd-a9c7-69e2-a568-75c304acd7c4, 'name': SearchDatastore_Task, 'duration_secs': 0.012256} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.203746] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f007856a-1ee6-407e-b83a-46194adcf086 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.209950] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1187.209950] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526b6aba-4580-836e-2cd1-6bf291bbd3bc" [ 1187.209950] env[65107]: _type = "Task" [ 1187.209950] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.220216] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526b6aba-4580-836e-2cd1-6bf291bbd3bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.428662] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "refresh_cache-10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.428662] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "refresh_cache-10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1187.428908] env[65107]: DEBUG nova.network.neutron [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1187.430280] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.430555] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.430690] env[65107]: INFO nova.compute.manager [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Shelving [ 1187.588591] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103737, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.709209] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.720770] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]526b6aba-4580-836e-2cd1-6bf291bbd3bc, 'name': SearchDatastore_Task, 'duration_secs': 0.010455} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.721046] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.721309] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 4a6894d7-233a-4c5c-8ded-5b80b9beadc7/4a6894d7-233a-4c5c-8ded-5b80b9beadc7.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1187.721615] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bcc4eb0-5a0a-4894-8735-81e223c54599 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.729488] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1187.729488] env[65107]: value = "task-5103738" [ 1187.729488] env[65107]: _type = "Task" [ 1187.729488] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.740576] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.860539] env[65107]: DEBUG nova.compute.manager [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Received event network-vif-plugged-3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1187.860892] env[65107]: DEBUG oslo_concurrency.lockutils [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Acquiring lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.861169] env[65107]: DEBUG oslo_concurrency.lockutils [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.861397] env[65107]: DEBUG oslo_concurrency.lockutils [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1187.861649] env[65107]: DEBUG nova.compute.manager [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] No waiting events found dispatching network-vif-plugged-3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1187.861887] env[65107]: WARNING nova.compute.manager [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Received unexpected event network-vif-plugged-3a1da1f7-19b4-4cb6-81d4-59775cea63dc for instance with vm_state building and task_state spawning. [ 1187.862112] env[65107]: DEBUG nova.compute.manager [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Received event network-changed-3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1187.862313] env[65107]: DEBUG nova.compute.manager [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Refreshing instance network info cache due to event network-changed-3a1da1f7-19b4-4cb6-81d4-59775cea63dc. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1187.862533] env[65107]: DEBUG oslo_concurrency.lockutils [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Acquiring lock "refresh_cache-10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.934181] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.934392] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.974905] env[65107]: DEBUG nova.network.neutron [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1187.997428] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.997828] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1188.074150] env[65107]: WARNING neutronclient.v2_0.client [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1188.074946] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1188.075391] env[65107]: WARNING openstack [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1188.097142] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103737, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.179023] env[65107]: DEBUG nova.network.neutron [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Updating instance_info_cache with network_info: [{"id": "3a1da1f7-19b4-4cb6-81d4-59775cea63dc", "address": "fa:16:3e:6e:f1:53", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1da1f7-19", "ovs_interfaceid": "3a1da1f7-19b4-4cb6-81d4-59775cea63dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1188.220011] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_power_states {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.241642] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504056} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.241976] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 4a6894d7-233a-4c5c-8ded-5b80b9beadc7/4a6894d7-233a-4c5c-8ded-5b80b9beadc7.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1188.242304] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1188.242560] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d313d141-d830-44c8-9e70-88f5c317b1a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.252298] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1188.252298] env[65107]: value = "task-5103739" [ 1188.252298] env[65107]: _type = "Task" [ 1188.252298] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.265430] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103739, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.451153] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.451556] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e667673-4418-410d-84a9-ab5fca7d18f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.459316] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1188.459316] env[65107]: value = "task-5103740" [ 1188.459316] env[65107]: _type = "Task" [ 1188.459316] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.470280] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.593514] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103737, 'name': CreateSnapshot_Task, 'duration_secs': 1.095357} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.593868] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1188.594583] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea834fc6-2fcd-4c58-987c-1a4aaee4d179 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.683022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "refresh_cache-10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1188.683450] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Instance network_info: |[{"id": "3a1da1f7-19b4-4cb6-81d4-59775cea63dc", "address": "fa:16:3e:6e:f1:53", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1da1f7-19", "ovs_interfaceid": "3a1da1f7-19b4-4cb6-81d4-59775cea63dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1188.683793] env[65107]: DEBUG oslo_concurrency.lockutils [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Acquired lock "refresh_cache-10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1188.683970] env[65107]: DEBUG nova.network.neutron [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Refreshing network info cache for port 3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1188.685344] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:f1:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a1da1f7-19b4-4cb6-81d4-59775cea63dc', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1188.692828] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1188.693327] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1188.693564] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-549d66fa-693e-4e3a-8f23-966c183372ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.714674] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1188.714674] env[65107]: value = "task-5103741" [ 1188.714674] env[65107]: _type = "Task" [ 1188.714674] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.723193] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103741, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.726966] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Getting list of instances from cluster (obj){ [ 1188.726966] env[65107]: value = "domain-c8" [ 1188.726966] env[65107]: _type = "ClusterComputeResource" [ 1188.726966] env[65107]: } {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1188.728192] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1705bb25-19b6-4cb5-8097-8c576be6fa8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.751990] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Got total of 11 instances {{(pid=65107) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1188.752196] env[65107]: WARNING nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] While synchronizing instance power states, found 12 instances in the database and 11 instances on the hypervisor. [ 1188.752343] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.752661] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 3df65e64-0c45-4707-960e-8f1767e2d011 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.752876] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 2c45d232-a948-4ad7-80a0-589c2dc91ec1 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.753103] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid d7b4bc1e-c94e-4654-9345-2a8aa945a896 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.753306] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid bef91b92-add1-4df9-bc24-dab15ce04338 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.753503] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.753700] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.753890] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.754101] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.754303] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.754531] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 4a6894d7-233a-4c5c-8ded-5b80b9beadc7 {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.754840] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Triggering sync for uuid 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad {{(pid=65107) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1188.755423] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.755638] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.755907] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "3df65e64-0c45-4707-960e-8f1767e2d011" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.756105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "3df65e64-0c45-4707-960e-8f1767e2d011" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.756356] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.756538] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.756793] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.756978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.757235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.757409] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.757642] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.757852] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.758041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.758285] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.758459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.758693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.758901] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.759086] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.759325] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.759775] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.764210] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9df44c-bab4-4eab-9773-3bf8d4a2ce3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.767341] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dea2aa6-ae42-412b-b11e-f7ee65606481 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.770478] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a3c940-9d67-499a-8831-cda136a42eae {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.774136] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caadc904-3080-45a4-870e-e7623e7b3e18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.777233] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d521e144-a1b4-4415-aa2f-3633beaf4659 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.781577] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69993910-3c10-4d4c-8669-7daa0f1452fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.785406] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24a4688-1cad-4ef6-b210-8109bbb48853 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.789594] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e1f70c-a8c1-481f-9ce0-01390ee95c61 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.818031] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103739, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06563} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.822722] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1188.831103] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef45615f-e863-423b-9095-226095b1585a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.863036] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 4a6894d7-233a-4c5c-8ded-5b80b9beadc7/4a6894d7-233a-4c5c-8ded-5b80b9beadc7.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.863253] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e48ac484-ddcf-4a24-8fe5-9705aa763714 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.887715] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1188.887715] env[65107]: value = "task-5103742" [ 1188.887715] env[65107]: _type = "Task" [ 1188.887715] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.897645] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103742, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.970420] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103740, 'name': PowerOffVM_Task, 'duration_secs': 0.451728} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.970709] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1188.971536] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63f52aa-0ecd-43c8-b0df-2d98097caf82 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.993066] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade80f2d-b493-486f-b123-eab914811d5f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.002474] env[65107]: WARNING oslo_messaging._drivers.amqpdriver [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 1189.113054] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1189.113419] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-46d29a12-f12a-4971-974d-51a203f35ee9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.123691] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1189.123691] env[65107]: value = "task-5103743" [ 1189.123691] env[65107]: _type = "Task" [ 1189.123691] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.132504] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103743, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.195026] env[65107]: WARNING neutronclient.v2_0.client [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.195945] env[65107]: WARNING openstack [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.196395] env[65107]: WARNING openstack [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.225084] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103741, 'name': CreateVM_Task, 'duration_secs': 0.366982} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.225289] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1189.225901] env[65107]: WARNING neutronclient.v2_0.client [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.226297] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.226453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.226787] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1189.227071] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c05caf2-e2ba-41e8-b369-82df2429980e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.232192] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1189.232192] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c21b11-2a9d-78ed-e0ef-d7ced665fe13" [ 1189.232192] env[65107]: _type = "Task" [ 1189.232192] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.243817] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c21b11-2a9d-78ed-e0ef-d7ced665fe13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.298604] env[65107]: WARNING openstack [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.298993] env[65107]: WARNING openstack [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.321509] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.564s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.322050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.565s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.327964] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.572s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.329757] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "3df65e64-0c45-4707-960e-8f1767e2d011" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.573s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.330315] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.572s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.338373] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.581s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.339073] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.581s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.365336] env[65107]: WARNING neutronclient.v2_0.client [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.366122] env[65107]: WARNING openstack [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.366480] env[65107]: WARNING openstack [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.381783] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.622s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.398230] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103742, 'name': ReconfigVM_Task, 'duration_secs': 0.363556} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.398888] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 4a6894d7-233a-4c5c-8ded-5b80b9beadc7/4a6894d7-233a-4c5c-8ded-5b80b9beadc7.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.399207] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-faee8815-bc7c-450d-b8bf-13f999303585 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.409218] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1189.409218] env[65107]: value = "task-5103744" [ 1189.409218] env[65107]: _type = "Task" [ 1189.409218] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.423678] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103744, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.460760] env[65107]: DEBUG nova.network.neutron [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Updated VIF entry in instance network info cache for port 3a1da1f7-19b4-4cb6-81d4-59775cea63dc. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1189.461235] env[65107]: DEBUG nova.network.neutron [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Updating instance_info_cache with network_info: [{"id": "3a1da1f7-19b4-4cb6-81d4-59775cea63dc", "address": "fa:16:3e:6e:f1:53", "network": {"id": "fe940fe7-c730-4320-94f7-30ed23a38c3c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1946195836-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d28d5a8ab6f747628edf3aee83355feb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1da1f7-19", "ovs_interfaceid": "3a1da1f7-19b4-4cb6-81d4-59775cea63dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1189.505796] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1189.506217] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a0667bf7-b79f-4275-9c06-f4987cad4891 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.516448] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1189.516448] env[65107]: value = "task-5103745" [ 1189.516448] env[65107]: _type = "Task" [ 1189.516448] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.526046] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103745, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.635767] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103743, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.743535] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c21b11-2a9d-78ed-e0ef-d7ced665fe13, 'name': SearchDatastore_Task, 'duration_secs': 0.014797} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.744315] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1189.744641] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1189.744931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.745137] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.745417] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1189.745741] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46f8c32f-16d5-48f7-a52b-d649f409fba5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.754774] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1189.755050] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1189.755753] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98071818-ace9-4d1d-a005-50cf32d3f016 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.761617] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1189.761617] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e9115e-4ab5-7eb5-8952-a9078f33d042" [ 1189.761617] env[65107]: _type = "Task" [ 1189.761617] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.770714] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e9115e-4ab5-7eb5-8952-a9078f33d042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.925054] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103744, 'name': Rename_Task, 'duration_secs': 0.17653} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.925312] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.925635] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-551f2a89-101d-4309-8ccd-af9bb93cbf3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.932380] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1189.932380] env[65107]: value = "task-5103746" [ 1189.932380] env[65107]: _type = "Task" [ 1189.932380] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.943146] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.964030] env[65107]: DEBUG oslo_concurrency.lockutils [req-f4e6f693-a22b-4840-96f2-25ff422b9d38 req-b43ac3e7-f4e5-418d-81c1-2525ce2800e9 service nova] Releasing lock "refresh_cache-10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1190.026978] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103745, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.135788] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103743, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.272544] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e9115e-4ab5-7eb5-8952-a9078f33d042, 'name': SearchDatastore_Task, 'duration_secs': 0.009624} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.273457] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23296c77-4b30-4d29-89e7-d30ae8bebc84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.280390] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1190.280390] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcc2f4-ef5f-edca-ff39-a5c31f7f47ff" [ 1190.280390] env[65107]: _type = "Task" [ 1190.280390] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.290480] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcc2f4-ef5f-edca-ff39-a5c31f7f47ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.442986] env[65107]: DEBUG oslo_vmware.api [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103746, 'name': PowerOnVM_Task, 'duration_secs': 0.463278} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.443185] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1190.443398] env[65107]: INFO nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Took 7.47 seconds to spawn the instance on the hypervisor. [ 1190.443577] env[65107]: DEBUG nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1190.444500] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf4622a-7c75-4e17-bb65-7bfebbb1d7dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.533516] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103745, 'name': CreateSnapshot_Task, 'duration_secs': 0.738572} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.533831] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1190.534677] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd52d253-2365-4679-b811-208babaa5210 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.636735] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103743, 'name': CloneVM_Task, 'duration_secs': 1.508564} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.636943] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Created linked-clone VM from snapshot [ 1190.637744] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c038adc4-1b8a-4502-97a3-7c8cb9352ec6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.646440] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Uploading image e4b7eab7-f146-4e21-9038-cf84fc47c977 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1190.673718] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1190.673718] env[65107]: value = "vm-992895" [ 1190.673718] env[65107]: _type = "VirtualMachine" [ 1190.673718] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1190.674036] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1cde9bc2-57a4-431d-baf9-dbbf13e72ac6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.682514] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease: (returnval){ [ 1190.682514] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207ee24-d0d3-3945-6bee-89d88e8bdfe4" [ 1190.682514] env[65107]: _type = "HttpNfcLease" [ 1190.682514] env[65107]: } obtained for exporting VM: (result){ [ 1190.682514] env[65107]: value = "vm-992895" [ 1190.682514] env[65107]: _type = "VirtualMachine" [ 1190.682514] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1190.682760] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the lease: (returnval){ [ 1190.682760] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207ee24-d0d3-3945-6bee-89d88e8bdfe4" [ 1190.682760] env[65107]: _type = "HttpNfcLease" [ 1190.682760] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1190.691757] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1190.691757] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207ee24-d0d3-3945-6bee-89d88e8bdfe4" [ 1190.691757] env[65107]: _type = "HttpNfcLease" [ 1190.691757] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1190.790632] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcc2f4-ef5f-edca-ff39-a5c31f7f47ff, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.790913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1190.791192] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad/10c2fd64-1f3b-41eb-a48c-34d4a892e2ad.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1190.791461] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7c6032a-8c64-4bf4-8699-9c4d6df47838 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.798594] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1190.798594] env[65107]: value = "task-5103748" [ 1190.798594] env[65107]: _type = "Task" [ 1190.798594] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.807712] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.965427] env[65107]: INFO nova.compute.manager [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Took 12.74 seconds to build instance. [ 1191.054196] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1191.054952] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-024d64ef-bb5d-4c03-8841-5089b431770a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.064969] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1191.064969] env[65107]: value = "task-5103749" [ 1191.064969] env[65107]: _type = "Task" [ 1191.064969] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.073602] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103749, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.195064] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1191.195064] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207ee24-d0d3-3945-6bee-89d88e8bdfe4" [ 1191.195064] env[65107]: _type = "HttpNfcLease" [ 1191.195064] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1191.195499] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1191.195499] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5207ee24-d0d3-3945-6bee-89d88e8bdfe4" [ 1191.195499] env[65107]: _type = "HttpNfcLease" [ 1191.195499] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1191.196266] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74513d66-39a2-400e-93b0-6666ac25ac2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.205452] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52883a77-0425-0d07-c065-f360eb84caac/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1191.205700] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52883a77-0425-0d07-c065-f360eb84caac/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1191.309020] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e075a1f4-e190-4535-b063-73064a39a114 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.315641] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103748, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.468111] env[65107]: DEBUG oslo_concurrency.lockutils [None req-866ab70f-6145-44f8-bbcf-182b7283168c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.257s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.468580] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.709s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.468788] env[65107]: INFO nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] During sync_power_state the instance has a pending task (spawning). Skip. [ 1191.468969] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.579678] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103749, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.811858] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.729335} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.811858] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad/10c2fd64-1f3b-41eb-a48c-34d4a892e2ad.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1191.812087] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1191.812377] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a39fc40-002b-4f93-a0bd-b89769e81ca8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.822447] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1191.822447] env[65107]: value = "task-5103750" [ 1191.822447] env[65107]: _type = "Task" [ 1191.822447] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.835667] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.040412] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1192.079544] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103749, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.202641] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1192.203377] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1192.203977] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1192.204402] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1192.204848] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.208061] env[65107]: INFO nova.compute.manager [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Terminating instance [ 1192.335690] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089553} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.336019] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1192.337016] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b66f657-28a5-4335-acd0-7f5322c7c2aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.367913] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad/10c2fd64-1f3b-41eb-a48c-34d4a892e2ad.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1192.368936] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f83ba2a4-e40e-4492-9d9f-dcb1fc8a633d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.390947] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1192.390947] env[65107]: value = "task-5103751" [ 1192.390947] env[65107]: _type = "Task" [ 1192.390947] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.401215] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103751, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.500255] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1192.500455] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1192.585240] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103749, 'name': CloneVM_Task} progress is 95%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.712867] env[65107]: DEBUG nova.compute.manager [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1192.713572] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1192.714707] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128daed6-6bd5-45fe-80ad-e0537b1c30da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.723654] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1192.724085] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-465d128a-6485-4fb5-8add-fa263552f74f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.732429] env[65107]: DEBUG oslo_vmware.api [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1192.732429] env[65107]: value = "task-5103752" [ 1192.732429] env[65107]: _type = "Task" [ 1192.732429] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.743697] env[65107]: DEBUG oslo_vmware.api [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.903943] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103751, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.078437] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103749, 'name': CloneVM_Task, 'duration_secs': 1.849582} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.078801] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Created linked-clone VM from snapshot [ 1193.079707] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb00027-b2c9-4278-815a-907cbe7ad577 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.088785] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Uploading image 830c17b6-eb60-4df2-8862-7545a6ab062f {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1193.116894] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1193.116894] env[65107]: value = "vm-992897" [ 1193.116894] env[65107]: _type = "VirtualMachine" [ 1193.116894] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1193.117322] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-02d99df8-2c6f-4e29-a3a5-beb66e6271f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.125298] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lease: (returnval){ [ 1193.125298] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849d1b-ec77-b2ac-a02f-cfe1a789dbd3" [ 1193.125298] env[65107]: _type = "HttpNfcLease" [ 1193.125298] env[65107]: } obtained for exporting VM: (result){ [ 1193.125298] env[65107]: value = "vm-992897" [ 1193.125298] env[65107]: _type = "VirtualMachine" [ 1193.125298] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1193.126266] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the lease: (returnval){ [ 1193.126266] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849d1b-ec77-b2ac-a02f-cfe1a789dbd3" [ 1193.126266] env[65107]: _type = "HttpNfcLease" [ 1193.126266] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1193.133163] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.133163] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849d1b-ec77-b2ac-a02f-cfe1a789dbd3" [ 1193.133163] env[65107]: _type = "HttpNfcLease" [ 1193.133163] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1193.243332] env[65107]: DEBUG oslo_vmware.api [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103752, 'name': PowerOffVM_Task, 'duration_secs': 0.200205} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.243653] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1193.243945] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1193.244255] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bdab0a7-9512-4e22-ac65-da2787bfcf16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.311597] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1193.311929] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1193.312223] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore2] 4a6894d7-233a-4c5c-8ded-5b80b9beadc7 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1193.312593] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bdc17b9-cee2-4ffa-b76a-3212356cd86e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.320953] env[65107]: DEBUG oslo_vmware.api [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1193.320953] env[65107]: value = "task-5103755" [ 1193.320953] env[65107]: _type = "Task" [ 1193.320953] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.330962] env[65107]: DEBUG oslo_vmware.api [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.402507] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103751, 'name': ReconfigVM_Task, 'duration_secs': 0.640664} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.402892] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Reconfigured VM instance instance-00000078 to attach disk [datastore2] 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad/10c2fd64-1f3b-41eb-a48c-34d4a892e2ad.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1193.403624] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-847e7152-24f3-48cc-8d3a-00ee98c8b38d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.412075] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1193.412075] env[65107]: value = "task-5103756" [ 1193.412075] env[65107]: _type = "Task" [ 1193.412075] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.420985] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103756, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.500908] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.501182] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.634530] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.634530] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849d1b-ec77-b2ac-a02f-cfe1a789dbd3" [ 1193.634530] env[65107]: _type = "HttpNfcLease" [ 1193.634530] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1193.635014] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1193.635014] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52849d1b-ec77-b2ac-a02f-cfe1a789dbd3" [ 1193.635014] env[65107]: _type = "HttpNfcLease" [ 1193.635014] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1193.635958] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a10434-1ec1-47ac-a2e7-ac5327fc4f48 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.644815] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf5fc5-e651-c8fa-d7df-5998169adc20/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1193.645082] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf5fc5-e651-c8fa-d7df-5998169adc20/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1193.763037] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f92fede6-eaec-492c-bcf2-f356b1be40ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.833116] env[65107]: DEBUG oslo_vmware.api [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148533} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.833460] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1193.833670] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1193.833869] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1193.834063] env[65107]: INFO nova.compute.manager [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1193.834374] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1193.834603] env[65107]: DEBUG nova.compute.manager [-] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1193.834719] env[65107]: DEBUG nova.network.neutron [-] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1193.834973] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.835580] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.835870] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.920513] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.929822] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103756, 'name': Rename_Task, 'duration_secs': 0.152352} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.930206] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1193.930509] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ef80fc9-e230-4f6e-8b5a-01ca562734ac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.937655] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1193.937655] env[65107]: value = "task-5103757" [ 1193.937655] env[65107]: _type = "Task" [ 1193.937655] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.947459] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103757, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.207869] env[65107]: DEBUG nova.compute.manager [req-1a02930b-67d1-48d0-ab00-293226d23992 req-7e25d75f-189d-4ddd-b308-3e3b22ff6b84 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Received event network-vif-deleted-c54e07bb-9d9b-4424-84ec-301f0abb7e4e {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1194.208176] env[65107]: INFO nova.compute.manager [req-1a02930b-67d1-48d0-ab00-293226d23992 req-7e25d75f-189d-4ddd-b308-3e3b22ff6b84 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Neutron deleted interface c54e07bb-9d9b-4424-84ec-301f0abb7e4e; detaching it from the instance and deleting it from the info cache [ 1194.208523] env[65107]: DEBUG nova.network.neutron [req-1a02930b-67d1-48d0-ab00-293226d23992 req-7e25d75f-189d-4ddd-b308-3e3b22ff6b84 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.450188] env[65107]: DEBUG oslo_vmware.api [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103757, 'name': PowerOnVM_Task, 'duration_secs': 0.498492} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.450678] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1194.450962] env[65107]: INFO nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Took 7.91 seconds to spawn the instance on the hypervisor. [ 1194.451214] env[65107]: DEBUG nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1194.452432] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d304e1bf-f4d0-4be0-b956-ebe2624852f4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.495128] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.678669] env[65107]: DEBUG nova.network.neutron [-] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.711988] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7eb6632-ff9c-4434-91a0-535cd6de3fff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.723308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f565b471-9680-4a75-adf6-d84a1d1a8c6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.764923] env[65107]: DEBUG nova.compute.manager [req-1a02930b-67d1-48d0-ab00-293226d23992 req-7e25d75f-189d-4ddd-b308-3e3b22ff6b84 service nova] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Detach interface failed, port_id=c54e07bb-9d9b-4424-84ec-301f0abb7e4e, reason: Instance 4a6894d7-233a-4c5c-8ded-5b80b9beadc7 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1194.974937] env[65107]: INFO nova.compute.manager [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Took 12.77 seconds to build instance. [ 1195.181363] env[65107]: INFO nova.compute.manager [-] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Took 1.35 seconds to deallocate network for instance. [ 1195.477726] env[65107]: DEBUG oslo_concurrency.lockutils [None req-72091f5a-e0ec-413e-82d2-f1d0b36c1d24 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.283s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.478213] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.718s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.478479] env[65107]: INFO nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] During sync_power_state the instance has a pending task (spawning). Skip. [ 1195.478672] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.500019] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.635672] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.635672] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.636148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.636148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.636148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.638564] env[65107]: INFO nova.compute.manager [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Terminating instance [ 1195.688862] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.689311] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.689427] env[65107]: DEBUG nova.objects.instance [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'resources' on Instance uuid 4a6894d7-233a-4c5c-8ded-5b80b9beadc7 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.143424] env[65107]: DEBUG nova.compute.manager [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1196.143751] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1196.145314] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf2ca36-0942-47ff-acb6-daa6c6542389 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.157251] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.157566] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40b97324-2a48-47f0-b41a-cf8fc17d4663 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.167252] env[65107]: DEBUG oslo_vmware.api [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1196.167252] env[65107]: value = "task-5103758" [ 1196.167252] env[65107]: _type = "Task" [ 1196.167252] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.177065] env[65107]: DEBUG oslo_vmware.api [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.400372] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210f016f-7324-46bf-9218-c315a01902ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.410756] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fc3b13-2a49-4529-a44a-d50fdb910da0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.445562] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f24897-c6c2-41c4-af90-5acff14782a8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.454673] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824472d6-0141-4b4b-a644-a135baa218e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.470802] env[65107]: DEBUG nova.compute.provider_tree [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.500211] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.679266] env[65107]: DEBUG oslo_vmware.api [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103758, 'name': PowerOffVM_Task, 'duration_secs': 0.213319} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.679627] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.679838] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.680147] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d15f3122-aeea-43ef-b3aa-1a956d4eb066 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.752644] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.753056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.753200] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleting the datastore file [datastore2] 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.753518] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34870f8b-123d-4a64-b83a-db30c54d2975 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.761669] env[65107]: DEBUG oslo_vmware.api [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for the task: (returnval){ [ 1196.761669] env[65107]: value = "task-5103760" [ 1196.761669] env[65107]: _type = "Task" [ 1196.761669] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.771822] env[65107]: DEBUG oslo_vmware.api [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.975418] env[65107]: DEBUG nova.scheduler.client.report [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1197.070618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1197.071833] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1197.273856] env[65107]: DEBUG oslo_vmware.api [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Task: {'id': task-5103760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176782} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.274161] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.274400] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1197.274619] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1197.274805] env[65107]: INFO nova.compute.manager [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1197.275074] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1197.275285] env[65107]: DEBUG nova.compute.manager [-] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1197.275420] env[65107]: DEBUG nova.network.neutron [-] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1197.275685] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.276255] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.276604] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.327815] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.482420] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.793s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.509530] env[65107]: INFO nova.scheduler.client.report [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance 4a6894d7-233a-4c5c-8ded-5b80b9beadc7 [ 1197.574413] env[65107]: DEBUG nova.compute.utils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1197.624158] env[65107]: DEBUG nova.compute.manager [req-8bcb068e-6505-4041-827e-51f6efcca745 req-e4bd463f-70fa-47f1-a586-78de685d3b73 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Received event network-vif-deleted-3a1da1f7-19b4-4cb6-81d4-59775cea63dc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1197.624465] env[65107]: INFO nova.compute.manager [req-8bcb068e-6505-4041-827e-51f6efcca745 req-e4bd463f-70fa-47f1-a586-78de685d3b73 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Neutron deleted interface 3a1da1f7-19b4-4cb6-81d4-59775cea63dc; detaching it from the instance and deleting it from the info cache [ 1197.624694] env[65107]: DEBUG nova.network.neutron [req-8bcb068e-6505-4041-827e-51f6efcca745 req-e4bd463f-70fa-47f1-a586-78de685d3b73 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1198.020807] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10263e68-0821-4152-910e-ff97b1041caa tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "4a6894d7-233a-4c5c-8ded-5b80b9beadc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.817s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.059557] env[65107]: DEBUG nova.network.neutron [-] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1198.079406] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.128020] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b43096b1-b9e9-4973-9b69-ef140774767c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.139896] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ebf403-1515-4abd-a544-95b3d0dc2323 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.179780] env[65107]: DEBUG nova.compute.manager [req-8bcb068e-6505-4041-827e-51f6efcca745 req-e4bd463f-70fa-47f1-a586-78de685d3b73 service nova] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Detach interface failed, port_id=3a1da1f7-19b4-4cb6-81d4-59775cea63dc, reason: Instance 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1198.563182] env[65107]: INFO nova.compute.manager [-] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Took 1.29 seconds to deallocate network for instance. [ 1199.070924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.071380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.071519] env[65107]: DEBUG nova.objects.instance [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lazy-loading 'resources' on Instance uuid 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.149787] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.150084] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.150330] env[65107]: INFO nova.compute.manager [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Attaching volume 226bb0c3-d519-44e2-bece-9023b17d955e to /dev/sdb [ 1199.186550] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1488a688-51d4-4a7d-a39e-bcdc4a5340ca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.194965] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fb3176-8fde-4d3d-8d42-a99908d50ae4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.209859] env[65107]: DEBUG nova.virt.block_device [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updating existing volume attachment record: 3d9a852e-9dc0-4516-bc51-ccc34fb25345 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1199.738109] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131fe584-865b-45bd-ab93-2b945f96d863 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.746675] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a364c04-b674-4f2c-8d4d-c7f0b9702ffc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.781791] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52883a77-0425-0d07-c065-f360eb84caac/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1199.782705] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfbb88d-d1d9-424b-b564-f434a72f6b47 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.790513] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d377dda6-bb62-4627-8748-5a5b24f42e8f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.803472] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52883a77-0425-0d07-c065-f360eb84caac/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1199.803472] env[65107]: ERROR oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52883a77-0425-0d07-c065-f360eb84caac/disk-0.vmdk due to incomplete transfer. [ 1199.804432] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61460b45-8d17-4763-961b-059d9ef6aecc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.809014] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1892b4e1-3a3d-4f4b-b3f4-d06349bf6883 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.830990] env[65107]: DEBUG nova.compute.provider_tree [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1199.832532] env[65107]: DEBUG oslo_vmware.rw_handles [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52883a77-0425-0d07-c065-f360eb84caac/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1199.832795] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Uploaded image e4b7eab7-f146-4e21-9038-cf84fc47c977 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1199.835583] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1199.836473] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f783bdbf-32a1-41d4-bcc9-2607c64e9dfa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.846448] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1199.846448] env[65107]: value = "task-5103762" [ 1199.846448] env[65107]: _type = "Task" [ 1199.846448] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.857952] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103762, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.294515] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.294855] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.295177] env[65107]: INFO nova.compute.manager [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Shelving [ 1200.337639] env[65107]: DEBUG nova.scheduler.client.report [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1200.357549] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103762, 'name': Destroy_Task, 'duration_secs': 0.357008} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.358609] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Destroyed the VM [ 1200.358906] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1200.359205] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-153dedc0-9bea-46fe-a113-ca1dabeca726 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.368243] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1200.368243] env[65107]: value = "task-5103763" [ 1200.368243] env[65107]: _type = "Task" [ 1200.368243] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.379164] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103763, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.500019] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.768641] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.768910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.843583] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.772s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.864232] env[65107]: INFO nova.scheduler.client.report [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Deleted allocations for instance 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad [ 1200.879591] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103763, 'name': RemoveSnapshot_Task, 'duration_secs': 0.368171} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.879891] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1200.880184] env[65107]: DEBUG nova.compute.manager [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1200.881074] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d86a5e-ebb8-4d7a-b6f5-5ec14bc67c26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.004124] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1201.004551] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.004831] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1201.005089] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1201.006203] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3547ed8-fd01-4f32-b427-5cad984e6c76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.019201] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d701006-ae7b-43f4-8ee2-adc9a95be02c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.035993] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e811f5-7d8f-4224-ac83-850eaa748d2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.043828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c84f325-4fa4-4e76-b032-ca4947bdf3de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.075331] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178146MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1201.075515] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1201.075745] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.271435] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1201.302605] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf5fc5-e651-c8fa-d7df-5998169adc20/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1201.303542] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead6d96e-d21a-4632-a5cb-a7374c47c351 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.306523] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1201.307423] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53478e3d-634f-4305-bd72-55db89db20cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.314988] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf5fc5-e651-c8fa-d7df-5998169adc20/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1201.315180] env[65107]: ERROR oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf5fc5-e651-c8fa-d7df-5998169adc20/disk-0.vmdk due to incomplete transfer. [ 1201.316430] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9c2a8262-7bae-4ae4-92c9-aba2242de090 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.319414] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1201.319414] env[65107]: value = "task-5103764" [ 1201.319414] env[65107]: _type = "Task" [ 1201.319414] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.327121] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cf5fc5-e651-c8fa-d7df-5998169adc20/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1201.327347] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Uploaded image 830c17b6-eb60-4df2-8862-7545a6ab062f to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1201.330054] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1201.333803] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8c4d8225-1ec4-454e-b16e-4d67e6217d2d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.335702] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.341172] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1201.341172] env[65107]: value = "task-5103765" [ 1201.341172] env[65107]: _type = "Task" [ 1201.341172] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.349900] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103765, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.375868] env[65107]: DEBUG oslo_concurrency.lockutils [None req-c687b698-a60b-4679-9671-befe6a4ebbf5 tempest-ImagesOneServerNegativeTestJSON-1910886864 tempest-ImagesOneServerNegativeTestJSON-1910886864-project-member] Lock "10c2fd64-1f3b-41eb-a48c-34d4a892e2ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.740s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1201.395036] env[65107]: INFO nova.compute.manager [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Shelve offloading [ 1201.796474] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1201.830209] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103764, 'name': PowerOffVM_Task, 'duration_secs': 0.275282} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.830573] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1201.831427] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61be334-edc9-4d05-8a3b-3b51a8c6e68c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.854161] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b83c51-f929-4288-97e4-e1a59a6d73d3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.862636] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103765, 'name': Destroy_Task, 'duration_secs': 0.3492} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.864659] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Destroyed the VM [ 1201.864911] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1201.867438] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5382252a-0cc6-4356-91b9-0279d23ca93a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.874435] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1201.874435] env[65107]: value = "task-5103767" [ 1201.874435] env[65107]: _type = "Task" [ 1201.874435] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.884209] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.899082] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1201.899459] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a02c23c8-220a-4d51-a3ce-b22eac6f51c5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.907545] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1201.907545] env[65107]: value = "task-5103768" [ 1201.907545] env[65107]: _type = "Task" [ 1201.907545] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.917986] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1201.918181] env[65107]: DEBUG nova.compute.manager [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1201.918974] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30b4f2d-eda4-4021-ac6e-37bbd7e9ac9a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.925248] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.925410] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1201.925583] env[65107]: DEBUG nova.network.neutron [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1202.105978] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.106150] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3df65e64-0c45-4707-960e-8f1767e2d011 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.106331] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 2c45d232-a948-4ad7-80a0-589c2dc91ec1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.106464] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance d7b4bc1e-c94e-4654-9345-2a8aa945a896 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.106669] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bef91b92-add1-4df9-bc24-dab15ce04338 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.106769] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 93a1b356-b6d5-435b-b30f-ea5ade031b7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.106923] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 475851f6-24a2-4f0e-9567-d2a23947a5e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.107071] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.107224] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance bcced66b-1464-4834-892b-3d0e652da31a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.107366] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 751ebb31-1a27-4630-9213-4312a6fa9298 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1202.370809] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Creating Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1202.371304] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-30573ee9-055f-4717-ab44-8bb51a0d556a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.381102] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1202.381102] env[65107]: value = "task-5103769" [ 1202.381102] env[65107]: _type = "Task" [ 1202.381102] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.388486] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103767, 'name': RemoveSnapshot_Task, 'duration_secs': 0.414098} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.389314] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1202.389662] env[65107]: DEBUG nova.compute.manager [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1202.390526] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138744f8-32aa-427b-89b8-5bef9d44aabf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.396702] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103769, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.428370] env[65107]: WARNING neutronclient.v2_0.client [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1202.429440] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.429761] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1202.584673] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.585080] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1202.610709] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 2c80c51e-dff3-4e56-bc13-32b792f67d97 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1202.611110] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1202.611239] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=100GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '10', 'num_vm_active': '9', 'num_task_None': '7', 'num_os_type_None': '10', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '2', 'io_workload': '0', 'num_proj_7784762e36ae4c6fb30012ec801c3b88': '2', 'num_proj_645feeb0c8524267bff18c5f37f5f011': '1', 'num_vm_shelved': '1', 'num_task_shelving_offloading': '1', 'num_proj_f379144b78764fe394039d87b043a946': '1', 'num_task_shelving': '1', 'num_proj_dced4f9aef2f49cf990203b693533aa1': '1', 'num_proj_544c649453844164bac53c98b342543f': '1', 'num_task_shelving_image_uploading': '1', 'num_proj_782397ddd6cb40298dcf7f0da7353d67': '1', 'num_proj_84acb61dcc764cd79ad2e5b1e2850d16': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1202.649083] env[65107]: WARNING neutronclient.v2_0.client [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1202.649724] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.650080] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1202.746070] env[65107]: DEBUG nova.network.neutron [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1202.770817] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d354408e-f5d4-437e-89de-9778d750d2d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.781624] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f7d17f-0f82-4a96-928d-3b4be38952d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.816275] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335fef10-fa1d-47ac-a35f-4a861b7428e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.826212] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f491dca7-a66b-4251-9156-d1cb3e75f08b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.842074] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.893656] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103769, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.908137] env[65107]: INFO nova.compute.manager [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Shelve offloading [ 1203.249361] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1203.249627] env[65107]: WARNING neutronclient.v2_0.client [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.250355] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.250855] env[65107]: WARNING openstack [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.258262] env[65107]: WARNING neutronclient.v2_0.client [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.347557] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1203.394783] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103769, 'name': CreateSnapshot_Task, 'duration_secs': 0.933656} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.395124] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Created Snapshot of the VM instance {{(pid=65107) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1203.395904] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3f2141-7434-4bd0-bdc4-180bf9385adb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.411873] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.412159] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70a19e27-a852-48b6-9dbb-bc1153178065 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.419983] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1203.419983] env[65107]: value = "task-5103770" [ 1203.419983] env[65107]: _type = "Task" [ 1203.419983] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.431858] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103770, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.671537] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1203.672505] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc66cde-eba3-43da-8a88-017140e3e3ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.678909] env[65107]: DEBUG nova.compute.manager [req-37207392-099a-48de-8f95-8a08500da19f req-c38fc3f1-aee5-41b0-9f63-4851135b353b service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-vif-unplugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1203.679282] env[65107]: DEBUG oslo_concurrency.lockutils [req-37207392-099a-48de-8f95-8a08500da19f req-c38fc3f1-aee5-41b0-9f63-4851135b353b service nova] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1203.680156] env[65107]: DEBUG oslo_concurrency.lockutils [req-37207392-099a-48de-8f95-8a08500da19f req-c38fc3f1-aee5-41b0-9f63-4851135b353b service nova] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.680156] env[65107]: DEBUG oslo_concurrency.lockutils [req-37207392-099a-48de-8f95-8a08500da19f req-c38fc3f1-aee5-41b0-9f63-4851135b353b service nova] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.680156] env[65107]: DEBUG nova.compute.manager [req-37207392-099a-48de-8f95-8a08500da19f req-c38fc3f1-aee5-41b0-9f63-4851135b353b service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] No waiting events found dispatching network-vif-unplugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1203.680156] env[65107]: WARNING nova.compute.manager [req-37207392-099a-48de-8f95-8a08500da19f req-c38fc3f1-aee5-41b0-9f63-4851135b353b service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received unexpected event network-vif-unplugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 for instance with vm_state shelved and task_state shelving_offloading. [ 1203.682776] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.683040] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63fcdf77-479a-4d8b-bf61-6ca10cdf787b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.751997] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1203.752204] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1203.752561] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleting the datastore file [datastore1] 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1203.752864] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d5c6836-c4bb-4459-af05-5a5afdada5de {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.761046] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1203.761046] env[65107]: value = "task-5103772" [ 1203.761046] env[65107]: _type = "Task" [ 1203.761046] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.762165] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1203.762422] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992898', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'name': 'volume-226bb0c3-d519-44e2-bece-9023b17d955e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3', 'attached_at': '', 'detached_at': '', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'serial': '226bb0c3-d519-44e2-bece-9023b17d955e'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1203.763324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d723700b-6b84-46f5-8c99-a694d5a5d4b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.774848] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103772, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.786865] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f587c1-910d-4d51-8667-aa08ad5a3f3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.814393] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-226bb0c3-d519-44e2-bece-9023b17d955e/volume-226bb0c3-d519-44e2-bece-9023b17d955e.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1203.815123] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d11cb9b-fba7-424b-9cbe-4a7a6923d122 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.835090] env[65107]: DEBUG oslo_vmware.api [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1203.835090] env[65107]: value = "task-5103773" [ 1203.835090] env[65107]: _type = "Task" [ 1203.835090] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.844171] env[65107]: DEBUG oslo_vmware.api [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103773, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.854254] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1203.854501] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.779s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.854852] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.059s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.856901] env[65107]: INFO nova.compute.claims [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1203.918976] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Creating linked-clone VM from snapshot {{(pid=65107) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1203.919683] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7552231b-adb9-4232-ae35-7026312eabd2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.935771] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1203.935771] env[65107]: value = "task-5103774" [ 1203.935771] env[65107]: _type = "Task" [ 1203.935771] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.936166] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1203.936354] env[65107]: DEBUG nova.compute.manager [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1203.937231] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855b7a42-7891-4818-9c89-7996bfbe7234 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.950320] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.950729] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1203.951017] env[65107]: DEBUG nova.network.neutron [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1203.952343] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103774, 'name': CloneVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.272840] env[65107]: DEBUG oslo_vmware.api [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103772, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1649} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.275195] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1204.275195] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1204.275195] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1204.297342] env[65107]: INFO nova.scheduler.client.report [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted allocations for instance 93a1b356-b6d5-435b-b30f-ea5ade031b7e [ 1204.348940] env[65107]: DEBUG oslo_vmware.api [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103773, 'name': ReconfigVM_Task, 'duration_secs': 0.454608} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.349265] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-226bb0c3-d519-44e2-bece-9023b17d955e/volume-226bb0c3-d519-44e2-bece-9023b17d955e.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1204.355012] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-101389eb-6ea9-445a-b230-58ccbe76fe9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.374357] env[65107]: DEBUG oslo_vmware.api [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1204.374357] env[65107]: value = "task-5103775" [ 1204.374357] env[65107]: _type = "Task" [ 1204.374357] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.385171] env[65107]: DEBUG oslo_vmware.api [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.448240] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103774, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.455073] env[65107]: WARNING neutronclient.v2_0.client [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1204.455073] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.455073] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.577824] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.578251] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.647481] env[65107]: WARNING neutronclient.v2_0.client [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1204.648165] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.648619] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.731141] env[65107]: DEBUG nova.network.neutron [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1204.801846] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.885499] env[65107]: DEBUG oslo_vmware.api [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103775, 'name': ReconfigVM_Task, 'duration_secs': 0.160008} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.885824] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992898', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'name': 'volume-226bb0c3-d519-44e2-bece-9023b17d955e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3', 'attached_at': '', 'detached_at': '', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'serial': '226bb0c3-d519-44e2-bece-9023b17d955e'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1204.950915] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103774, 'name': CloneVM_Task} progress is 94%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.026863] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913d9460-103d-429e-ad35-9b9d2580691c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.034235] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb4c5b5-f1fa-43b4-ab02-8ee69dbf3c92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.066845] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eaaaf22-5856-47d6-98fe-73ff8ff36123 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.075132] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dce7b8-170b-42b7-a8a0-25a3a169eacb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.091038] env[65107]: DEBUG nova.compute.provider_tree [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.234605] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1205.235171] env[65107]: WARNING neutronclient.v2_0.client [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.236068] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.236554] env[65107]: WARNING openstack [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.244407] env[65107]: WARNING neutronclient.v2_0.client [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.449821] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103774, 'name': CloneVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.564484] env[65107]: DEBUG nova.compute.manager [req-8df32267-b36e-4934-9a07-40758da71356 req-745951e0-9041-4eb5-845c-7a45db3c06a2 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-vif-unplugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1205.564484] env[65107]: DEBUG oslo_concurrency.lockutils [req-8df32267-b36e-4934-9a07-40758da71356 req-745951e0-9041-4eb5-845c-7a45db3c06a2 service nova] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.564607] env[65107]: DEBUG oslo_concurrency.lockutils [req-8df32267-b36e-4934-9a07-40758da71356 req-745951e0-9041-4eb5-845c-7a45db3c06a2 service nova] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.564776] env[65107]: DEBUG oslo_concurrency.lockutils [req-8df32267-b36e-4934-9a07-40758da71356 req-745951e0-9041-4eb5-845c-7a45db3c06a2 service nova] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.564970] env[65107]: DEBUG nova.compute.manager [req-8df32267-b36e-4934-9a07-40758da71356 req-745951e0-9041-4eb5-845c-7a45db3c06a2 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] No waiting events found dispatching network-vif-unplugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1205.565160] env[65107]: WARNING nova.compute.manager [req-8df32267-b36e-4934-9a07-40758da71356 req-745951e0-9041-4eb5-845c-7a45db3c06a2 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received unexpected event network-vif-unplugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede for instance with vm_state shelved and task_state shelving_offloading. [ 1205.566425] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1205.568033] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9c7203-027d-43a1-9a75-10588e2ebedf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.576893] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1205.577184] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8c3b60d-7c27-43a0-b3b9-ab3e81d308b0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.594213] env[65107]: DEBUG nova.scheduler.client.report [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1205.658029] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1205.658029] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1205.658029] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleting the datastore file [datastore2] bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.658029] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5627302d-abaf-4803-9b3a-c24ecc7f1119 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.666484] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1205.666484] env[65107]: value = "task-5103777" [ 1205.666484] env[65107]: _type = "Task" [ 1205.666484] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.676234] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103777, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.710881] env[65107]: DEBUG nova.compute.manager [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1205.711052] env[65107]: DEBUG nova.compute.manager [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing instance network info cache due to event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1205.711284] env[65107]: DEBUG oslo_concurrency.lockutils [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.711581] env[65107]: DEBUG oslo_concurrency.lockutils [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.711790] env[65107]: DEBUG nova.network.neutron [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1205.927941] env[65107]: DEBUG nova.objects.instance [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'flavor' on Instance uuid 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.951600] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103774, 'name': CloneVM_Task, 'duration_secs': 1.561041} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.951916] env[65107]: INFO nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Created linked-clone VM from snapshot [ 1205.952687] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9a8b82-cf14-4a66-b3cd-4c3d4d6ba01c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.960471] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Uploading image 088b6ede-5bb4-4485-8338-801656621e71 {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1205.985224] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1205.985224] env[65107]: value = "vm-992900" [ 1205.985224] env[65107]: _type = "VirtualMachine" [ 1205.985224] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1205.985571] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b3dc6feb-8baf-45ac-8a9a-d507c264aa67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.993034] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease: (returnval){ [ 1205.993034] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c142b-2f7f-607a-5ba0-35019c6508b7" [ 1205.993034] env[65107]: _type = "HttpNfcLease" [ 1205.993034] env[65107]: } obtained for exporting VM: (result){ [ 1205.993034] env[65107]: value = "vm-992900" [ 1205.993034] env[65107]: _type = "VirtualMachine" [ 1205.993034] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1205.993034] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the lease: (returnval){ [ 1205.993034] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c142b-2f7f-607a-5ba0-35019c6508b7" [ 1205.993034] env[65107]: _type = "HttpNfcLease" [ 1205.993034] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1206.002750] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1206.002750] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c142b-2f7f-607a-5ba0-35019c6508b7" [ 1206.002750] env[65107]: _type = "HttpNfcLease" [ 1206.002750] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1206.068291] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.099080] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.244s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.099649] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1206.102340] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.301s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.102596] env[65107]: DEBUG nova.objects.instance [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'resources' on Instance uuid 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.178171] env[65107]: DEBUG oslo_vmware.api [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103777, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265747} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.178445] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.178628] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1206.178799] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1206.197638] env[65107]: INFO nova.scheduler.client.report [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted allocations for instance bcced66b-1464-4834-892b-3d0e652da31a [ 1206.214669] env[65107]: WARNING neutronclient.v2_0.client [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.215492] env[65107]: WARNING openstack [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.215852] env[65107]: WARNING openstack [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.347713] env[65107]: WARNING openstack [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.348094] env[65107]: WARNING openstack [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.409653] env[65107]: WARNING neutronclient.v2_0.client [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.410434] env[65107]: WARNING openstack [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.410843] env[65107]: WARNING openstack [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.433067] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3f2a300a-8e3a-4359-b9c0-7ada80c1a4a0 tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.283s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.493034] env[65107]: DEBUG nova.network.neutron [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updated VIF entry in instance network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1206.493414] env[65107]: DEBUG nova.network.neutron [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1206.502834] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1206.502834] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c142b-2f7f-607a-5ba0-35019c6508b7" [ 1206.502834] env[65107]: _type = "HttpNfcLease" [ 1206.502834] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1206.503229] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1206.503229] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c142b-2f7f-607a-5ba0-35019c6508b7" [ 1206.503229] env[65107]: _type = "HttpNfcLease" [ 1206.503229] env[65107]: }. {{(pid=65107) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1206.504133] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef839a1-9d8d-4066-b2cb-71fe648ab647 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.513068] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52073a71-7914-c95d-531e-11974c394272/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1206.513255] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52073a71-7914-c95d-531e-11974c394272/disk-0.vmdk for reading. {{(pid=65107) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1206.605687] env[65107]: DEBUG nova.compute.utils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1206.608217] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1206.608538] env[65107]: DEBUG nova.network.neutron [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1206.608962] env[65107]: WARNING neutronclient.v2_0.client [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.609434] env[65107]: WARNING neutronclient.v2_0.client [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.610129] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.610545] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.618227] env[65107]: DEBUG nova.objects.instance [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'numa_topology' on Instance uuid 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.622449] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dd3c9a59-2703-4601-950c-4a1ca68bae3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.664395] env[65107]: DEBUG nova.policy [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e787eb2cbd6454a9d1dd7a39e438092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '645feeb0c8524267bff18c5f37f5f011', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1206.666924] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.667227] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.702643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.986840] env[65107]: DEBUG nova.network.neutron [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Successfully created port: 6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1206.999335] env[65107]: DEBUG oslo_concurrency.lockutils [req-a0c6a2ef-579f-400c-b797-e43b841e7772 req-5bd4b4b8-13a3-4b8d-89a0-c7567c4db3ef service nova] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1207.108936] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1207.124867] env[65107]: DEBUG nova.objects.base [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Object Instance<93a1b356-b6d5-435b-b30f-ea5ade031b7e> lazy-loaded attributes: resources,numa_topology {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1207.171072] env[65107]: INFO nova.compute.manager [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Detaching volume 226bb0c3-d519-44e2-bece-9023b17d955e [ 1207.219228] env[65107]: INFO nova.virt.block_device [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Attempting to driver detach volume 226bb0c3-d519-44e2-bece-9023b17d955e from mountpoint /dev/sdb [ 1207.219228] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1207.219228] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992898', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'name': 'volume-226bb0c3-d519-44e2-bece-9023b17d955e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3', 'attached_at': '', 'detached_at': '', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'serial': '226bb0c3-d519-44e2-bece-9023b17d955e'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1207.220362] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298de580-42d6-47f1-8228-0a2b23de9263 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.257678] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c38b1d-6885-4a10-844f-93fbfdf5da1c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.266036] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec0fffc-9809-4f52-827d-9ed97bd7e500 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.293820] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2716595-2cfb-4eb7-b30c-957b4621a546 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.311255] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] The volume has not been displaced from its original location: [datastore1] volume-226bb0c3-d519-44e2-bece-9023b17d955e/volume-226bb0c3-d519-44e2-bece-9023b17d955e.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1207.316779] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1207.319888] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25ba3e62-49c1-47db-9f77-e4f291dfd596 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.342011] env[65107]: DEBUG oslo_vmware.api [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1207.342011] env[65107]: value = "task-5103779" [ 1207.342011] env[65107]: _type = "Task" [ 1207.342011] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.354575] env[65107]: DEBUG oslo_vmware.api [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103779, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.383033] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b666d2fb-62a1-459d-920d-790b08777338 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.390328] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f2c65b-08b9-433d-95eb-a89844592110 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.424198] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab621735-e774-47e5-b300-a2fd26fecaaf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.433807] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df72815-deb9-46b2-aa2c-62f7be3c0ecd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.452662] env[65107]: DEBUG nova.compute.provider_tree [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.593786] env[65107]: DEBUG nova.compute.manager [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-changed-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1207.593786] env[65107]: DEBUG nova.compute.manager [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Refreshing instance network info cache due to event network-changed-e48b19c2-6253-4a9c-a6fb-1daae3460ede. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1207.594107] env[65107]: DEBUG oslo_concurrency.lockutils [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.594107] env[65107]: DEBUG oslo_concurrency.lockutils [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1207.594248] env[65107]: DEBUG nova.network.neutron [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Refreshing network info cache for port e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1207.852709] env[65107]: DEBUG oslo_vmware.api [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103779, 'name': ReconfigVM_Task, 'duration_secs': 0.25671} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.853687] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.859253] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57bfb37c-1b1b-44b1-bfd8-7f702347cd16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.876937] env[65107]: DEBUG oslo_vmware.api [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1207.876937] env[65107]: value = "task-5103780" [ 1207.876937] env[65107]: _type = "Task" [ 1207.876937] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.886063] env[65107]: DEBUG oslo_vmware.api [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103780, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.955968] env[65107]: DEBUG nova.scheduler.client.report [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1208.098862] env[65107]: WARNING neutronclient.v2_0.client [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.099660] env[65107]: WARNING openstack [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.100171] env[65107]: WARNING openstack [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.118101] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1208.143583] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1208.143924] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1208.144157] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1208.144463] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1208.144710] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1208.144860] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1208.145093] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1208.145262] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1208.145495] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1208.145746] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1208.145976] env[65107]: DEBUG nova.virt.hardware [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1208.147169] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11699b1-abec-44aa-878d-443628673e99 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.157049] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13c25e0-c84c-41b3-877c-b0b056b7468a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.205462] env[65107]: WARNING openstack [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.205988] env[65107]: WARNING openstack [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.306458] env[65107]: WARNING neutronclient.v2_0.client [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.307340] env[65107]: WARNING openstack [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.307782] env[65107]: WARNING openstack [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.396718] env[65107]: DEBUG oslo_vmware.api [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103780, 'name': ReconfigVM_Task, 'duration_secs': 0.165022} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.397196] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992898', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'name': 'volume-226bb0c3-d519-44e2-bece-9023b17d955e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3', 'attached_at': '', 'detached_at': '', 'volume_id': '226bb0c3-d519-44e2-bece-9023b17d955e', 'serial': '226bb0c3-d519-44e2-bece-9023b17d955e'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1208.444674] env[65107]: DEBUG nova.network.neutron [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updated VIF entry in instance network info cache for port e48b19c2-6253-4a9c-a6fb-1daae3460ede. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1208.445183] env[65107]: DEBUG nova.network.neutron [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tape48b19c2-62", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1208.462072] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.360s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1208.465900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.763s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1208.466156] env[65107]: DEBUG nova.objects.instance [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'resources' on Instance uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.517978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1208.560142] env[65107]: DEBUG nova.network.neutron [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Successfully updated port: 6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1208.943130] env[65107]: DEBUG nova.objects.instance [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'flavor' on Instance uuid 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.947992] env[65107]: DEBUG oslo_concurrency.lockutils [req-a9427b7c-f93f-4caf-bdb6-52ff0d9d1577 req-873d6ff9-8046-424a-82e5-87e9c0416083 service nova] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1208.970905] env[65107]: DEBUG nova.objects.instance [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'numa_topology' on Instance uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.972827] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f282cf10-8dab-4bab-9247-1a0d6c802b69 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.988s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1208.973645] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.216s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1208.973824] env[65107]: INFO nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] During sync_power_state the instance has a pending task (shelving_image_pending_upload). Skip. [ 1208.973997] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1208.974482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.906s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1208.974787] env[65107]: INFO nova.compute.manager [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Unshelving [ 1209.063273] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "refresh_cache-2c80c51e-dff3-4e56-bc13-32b792f67d97" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.063273] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "refresh_cache-2c80c51e-dff3-4e56-bc13-32b792f67d97" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1209.063412] env[65107]: DEBUG nova.network.neutron [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1209.473431] env[65107]: DEBUG nova.objects.base [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1209.566583] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.567008] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.604119] env[65107]: DEBUG nova.network.neutron [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1209.624240] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.624635] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.635249] env[65107]: DEBUG nova.compute.manager [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Received event network-vif-plugged-6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1209.635493] env[65107]: DEBUG oslo_concurrency.lockutils [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Acquiring lock "2c80c51e-dff3-4e56-bc13-32b792f67d97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.635749] env[65107]: DEBUG oslo_concurrency.lockutils [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.635872] env[65107]: DEBUG oslo_concurrency.lockutils [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.636053] env[65107]: DEBUG nova.compute.manager [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] No waiting events found dispatching network-vif-plugged-6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1209.636219] env[65107]: WARNING nova.compute.manager [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Received unexpected event network-vif-plugged-6af11f76-b4de-4137-b77f-a285924025d9 for instance with vm_state building and task_state spawning. [ 1209.636375] env[65107]: DEBUG nova.compute.manager [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Received event network-changed-6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1209.636525] env[65107]: DEBUG nova.compute.manager [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Refreshing instance network info cache due to event network-changed-6af11f76-b4de-4137-b77f-a285924025d9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1209.636687] env[65107]: DEBUG oslo_concurrency.lockutils [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Acquiring lock "refresh_cache-2c80c51e-dff3-4e56-bc13-32b792f67d97" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.661775] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9f2084-c5c3-45f7-9727-0573ca8ac3d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.670280] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a108384-e39a-4a13-9a77-03ee525400e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.706955] env[65107]: WARNING neutronclient.v2_0.client [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.707626] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.707969] env[65107]: WARNING openstack [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.716187] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea560c7-df7b-4308-b2ef-4abb11b1bc4f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.724964] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92e43f8-3a8e-417a-8259-6da55383cdbc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.739893] env[65107]: DEBUG nova.compute.provider_tree [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.793696] env[65107]: DEBUG nova.network.neutron [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Updating instance_info_cache with network_info: [{"id": "6af11f76-b4de-4137-b77f-a285924025d9", "address": "fa:16:3e:84:7f:36", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6af11f76-b4", "ovs_interfaceid": "6af11f76-b4de-4137-b77f-a285924025d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1209.951782] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0347e25c-bf9c-46a3-aaf0-e86ad2fc78de tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.284s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.999962] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.243146] env[65107]: DEBUG nova.scheduler.client.report [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1210.296764] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "refresh_cache-2c80c51e-dff3-4e56-bc13-32b792f67d97" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1210.297184] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Instance network_info: |[{"id": "6af11f76-b4de-4137-b77f-a285924025d9", "address": "fa:16:3e:84:7f:36", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6af11f76-b4", "ovs_interfaceid": "6af11f76-b4de-4137-b77f-a285924025d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1210.297532] env[65107]: DEBUG oslo_concurrency.lockutils [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Acquired lock "refresh_cache-2c80c51e-dff3-4e56-bc13-32b792f67d97" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1210.297723] env[65107]: DEBUG nova.network.neutron [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Refreshing network info cache for port 6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1210.298927] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:7f:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6af11f76-b4de-4137-b77f-a285924025d9', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1210.306975] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1210.308573] env[65107]: WARNING neutronclient.v2_0.client [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.309339] env[65107]: WARNING openstack [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.309775] env[65107]: WARNING openstack [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.317330] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1210.318145] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cb82224-fdb0-46f5-9077-548b99f2bc4e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.339689] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1210.339689] env[65107]: value = "task-5103781" [ 1210.339689] env[65107]: _type = "Task" [ 1210.339689] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.348460] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103781, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.421355] env[65107]: WARNING openstack [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.421793] env[65107]: WARNING openstack [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.498228] env[65107]: WARNING neutronclient.v2_0.client [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.498956] env[65107]: WARNING openstack [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.499393] env[65107]: WARNING openstack [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.587039] env[65107]: DEBUG nova.network.neutron [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Updated VIF entry in instance network info cache for port 6af11f76-b4de-4137-b77f-a285924025d9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1210.587470] env[65107]: DEBUG nova.network.neutron [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Updating instance_info_cache with network_info: [{"id": "6af11f76-b4de-4137-b77f-a285924025d9", "address": "fa:16:3e:84:7f:36", "network": {"id": "c1832506-5268-44c8-8f86-f86261ad993f", "bridge": "br-int", "label": "tempest-ServersTestJSON-1974068052-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "645feeb0c8524267bff18c5f37f5f011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9079d3b9-5c2d-4ca1-8d2f-68ceb8ec8c98", "external-id": "nsx-vlan-transportzone-527", "segmentation_id": 527, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6af11f76-b4", "ovs_interfaceid": "6af11f76-b4de-4137-b77f-a285924025d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1210.748642] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.283s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.751189] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.751s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.751434] env[65107]: DEBUG nova.objects.instance [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'pci_requests' on Instance uuid 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1210.851126] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103781, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.016601] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.017034] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.017114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.017302] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.017469] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.019818] env[65107]: INFO nova.compute.manager [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Terminating instance [ 1211.090848] env[65107]: DEBUG oslo_concurrency.lockutils [req-32526ebc-2593-462b-98d3-9ee9594681e8 req-72089921-1a93-439e-864c-66f82de6ca42 service nova] Releasing lock "refresh_cache-2c80c51e-dff3-4e56-bc13-32b792f67d97" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1211.256083] env[65107]: DEBUG nova.objects.instance [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'numa_topology' on Instance uuid 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.261278] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b3879524-dac8-4feb-8047-e91486be5ec8 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.831s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.262259] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "bcced66b-1464-4834-892b-3d0e652da31a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 22.503s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.262481] env[65107]: INFO nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: bcced66b-1464-4834-892b-3d0e652da31a] During sync_power_state the instance has a pending task (shelving). Skip. [ 1211.262722] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "bcced66b-1464-4834-892b-3d0e652da31a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.263105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.745s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.263283] env[65107]: INFO nova.compute.manager [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Unshelving [ 1211.351339] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103781, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.523235] env[65107]: DEBUG nova.compute.manager [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1211.523428] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1211.524389] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e92f4a-d00c-497c-bdc9-2f7607ac6869 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.533277] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1211.533557] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-163ab325-dafd-41cf-af17-4a60dcb8e3b1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.540267] env[65107]: DEBUG oslo_vmware.api [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1211.540267] env[65107]: value = "task-5103782" [ 1211.540267] env[65107]: _type = "Task" [ 1211.540267] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.549361] env[65107]: DEBUG oslo_vmware.api [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.759600] env[65107]: INFO nova.compute.claims [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1211.851540] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103781, 'name': CreateVM_Task, 'duration_secs': 1.209444} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.851726] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1211.852315] env[65107]: WARNING neutronclient.v2_0.client [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.861279] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.861453] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1211.861772] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1211.862052] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbde1612-d7d3-45f5-8b1f-5238dac125c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.867843] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1211.867843] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52362459-85af-0903-1644-8ff70d363502" [ 1211.867843] env[65107]: _type = "Task" [ 1211.867843] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.877298] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52362459-85af-0903-1644-8ff70d363502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.053222] env[65107]: DEBUG oslo_vmware.api [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103782, 'name': PowerOffVM_Task, 'duration_secs': 0.20796} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.053617] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1212.053774] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1212.054084] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8cc08e0a-1171-4a65-9690-e56db9802d76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.132805] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1212.133046] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1212.133271] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleting the datastore file [datastore2] 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1212.133850] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-952516a3-d25c-4210-a74f-19494d75bad5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.142415] env[65107]: DEBUG oslo_vmware.api [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for the task: (returnval){ [ 1212.142415] env[65107]: value = "task-5103784" [ 1212.142415] env[65107]: _type = "Task" [ 1212.142415] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.154218] env[65107]: DEBUG oslo_vmware.api [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103784, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.301133] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.380179] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52362459-85af-0903-1644-8ff70d363502, 'name': SearchDatastore_Task, 'duration_secs': 0.010865} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.380543] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.380798] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1212.381073] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.381239] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.381437] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1212.381728] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cce9fafa-a355-475c-b918-6bd9b80817ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.402530] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1212.402786] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1212.403558] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45485c38-3faa-4390-9928-4dc3477c1cfb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.410328] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1212.410328] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b40fd7-2cc9-c5a9-34c2-c4aa8737b678" [ 1212.410328] env[65107]: _type = "Task" [ 1212.410328] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.419445] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b40fd7-2cc9-c5a9-34c2-c4aa8737b678, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.653067] env[65107]: DEBUG oslo_vmware.api [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Task: {'id': task-5103784, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157534} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.653333] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1212.653520] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1212.653716] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1212.653922] env[65107]: INFO nova.compute.manager [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1212.654189] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1212.654393] env[65107]: DEBUG nova.compute.manager [-] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1212.654499] env[65107]: DEBUG nova.network.neutron [-] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1212.654737] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.655289] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.655584] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.694551] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.924674] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b40fd7-2cc9-c5a9-34c2-c4aa8737b678, 'name': SearchDatastore_Task, 'duration_secs': 0.010481} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.928886] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aaa2fd1-38ee-4773-8c7f-23820d6057a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.936548] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1212.936548] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2f5d2-75b4-0538-c985-9b967d1bc8bc" [ 1212.936548] env[65107]: _type = "Task" [ 1212.936548] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.943732] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f4e34f-acc2-48fd-9341-86684f62fe42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.950129] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2f5d2-75b4-0538-c985-9b967d1bc8bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.955786] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa62735-41c0-40f2-bc7b-a1ffc24e2f2a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.991081] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f672cfd-7728-4d89-bac9-cd6c6230516d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.001130] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dabcca6-c0ef-49cc-83da-847c0fe039f1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.016774] env[65107]: DEBUG nova.compute.provider_tree [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.103527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.103527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.103527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.103527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.103527] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1213.105777] env[65107]: INFO nova.compute.manager [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Terminating instance [ 1213.168235] env[65107]: DEBUG nova.compute.manager [req-a1eb7cf9-7df5-455d-b108-e6e710ee38f4 req-fe7dd2f4-6ce8-43c2-b4f4-9d217bdb1608 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Received event network-vif-deleted-3914e94e-1c0c-410a-adcb-0ac7cb7ff3be {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1213.168474] env[65107]: INFO nova.compute.manager [req-a1eb7cf9-7df5-455d-b108-e6e710ee38f4 req-fe7dd2f4-6ce8-43c2-b4f4-9d217bdb1608 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Neutron deleted interface 3914e94e-1c0c-410a-adcb-0ac7cb7ff3be; detaching it from the instance and deleting it from the info cache [ 1213.168665] env[65107]: DEBUG nova.network.neutron [req-a1eb7cf9-7df5-455d-b108-e6e710ee38f4 req-fe7dd2f4-6ce8-43c2-b4f4-9d217bdb1608 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.452114] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f2f5d2-75b4-0538-c985-9b967d1bc8bc, 'name': SearchDatastore_Task, 'duration_secs': 0.014947} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.452114] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.452114] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2c80c51e-dff3-4e56-bc13-32b792f67d97/2c80c51e-dff3-4e56-bc13-32b792f67d97.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1213.452114] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-403cae26-405f-4bba-8a79-e6e1e95ae3c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.459877] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1213.459877] env[65107]: value = "task-5103785" [ 1213.459877] env[65107]: _type = "Task" [ 1213.459877] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.469209] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.520928] env[65107]: DEBUG nova.scheduler.client.report [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1213.612024] env[65107]: DEBUG nova.compute.manager [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1213.612024] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1213.612024] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43857b0a-7f4d-471c-968e-10571663cc6a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.621693] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1213.622268] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-484e882f-33bb-410f-a2d5-9a8755d0d2ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.630725] env[65107]: DEBUG oslo_vmware.api [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1213.630725] env[65107]: value = "task-5103786" [ 1213.630725] env[65107]: _type = "Task" [ 1213.630725] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.644043] env[65107]: DEBUG nova.network.neutron [-] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.645834] env[65107]: DEBUG oslo_vmware.api [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103786, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.671825] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5905a9d9-b137-4ef6-8095-70d51d704df0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.687194] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f8205f-b9c2-405c-bf0c-b3081b16d393 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.731440] env[65107]: DEBUG nova.compute.manager [req-a1eb7cf9-7df5-455d-b108-e6e710ee38f4 req-fe7dd2f4-6ce8-43c2-b4f4-9d217bdb1608 service nova] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Detach interface failed, port_id=3914e94e-1c0c-410a-adcb-0ac7cb7ff3be, reason: Instance 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1213.972530] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103785, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.028840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.277s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.029302] env[65107]: WARNING neutronclient.v2_0.client [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.032753] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.732s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1214.033075] env[65107]: DEBUG nova.objects.instance [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'pci_requests' on Instance uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.065586] env[65107]: INFO nova.network.neutron [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1214.126330] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1214.126330] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1214.143541] env[65107]: DEBUG oslo_vmware.api [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103786, 'name': PowerOffVM_Task, 'duration_secs': 0.251147} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.143624] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1214.143759] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1214.144051] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e879cfd8-23d3-4d40-8e89-28d835a4e8b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.148011] env[65107]: INFO nova.compute.manager [-] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Took 1.49 seconds to deallocate network for instance. [ 1214.217268] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1214.217657] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1214.217938] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleting the datastore file [datastore2] d7b4bc1e-c94e-4654-9345-2a8aa945a896 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1214.218335] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-691f14d1-98bf-4692-8903-37164e22a4b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.227683] env[65107]: DEBUG oslo_vmware.api [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1214.227683] env[65107]: value = "task-5103788" [ 1214.227683] env[65107]: _type = "Task" [ 1214.227683] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.238878] env[65107]: DEBUG oslo_vmware.api [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103788, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.471590] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586324} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.471796] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2c80c51e-dff3-4e56-bc13-32b792f67d97/2c80c51e-dff3-4e56-bc13-32b792f67d97.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1214.472040] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1214.472339] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1725256f-a372-4d6e-b4ec-2e2c5962e1c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.480303] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1214.480303] env[65107]: value = "task-5103789" [ 1214.480303] env[65107]: _type = "Task" [ 1214.480303] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.491481] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.540247] env[65107]: DEBUG nova.objects.instance [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'numa_topology' on Instance uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.630056] env[65107]: DEBUG nova.compute.utils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1214.655305] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1214.737954] env[65107]: DEBUG oslo_vmware.api [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103788, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152944} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.738246] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1214.738462] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1214.738684] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1214.738906] env[65107]: INFO nova.compute.manager [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1214.739183] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1214.739389] env[65107]: DEBUG nova.compute.manager [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1214.739488] env[65107]: DEBUG nova.network.neutron [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1214.739722] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.740258] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.740511] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.776677] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.990557] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086803} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.990827] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1214.991670] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf0c0cd-c0f2-4a01-ad21-8cb236eb6257 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.016655] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] 2c80c51e-dff3-4e56-bc13-32b792f67d97/2c80c51e-dff3-4e56-bc13-32b792f67d97.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1215.016655] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bcd5ead-ddd6-441b-ba37-8bfe010cb7c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.037399] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1215.037399] env[65107]: value = "task-5103790" [ 1215.037399] env[65107]: _type = "Task" [ 1215.037399] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.047297] env[65107]: INFO nova.compute.claims [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1215.050348] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103790, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.133028] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.142453] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52073a71-7914-c95d-531e-11974c394272/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1215.143486] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867bab89-7aa3-46d1-83d3-27d4494ba2c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.150102] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52073a71-7914-c95d-531e-11974c394272/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1215.150228] env[65107]: ERROR oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52073a71-7914-c95d-531e-11974c394272/disk-0.vmdk due to incomplete transfer. [ 1215.150391] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fb401dec-0a18-4196-aef4-0e0eb56e6210 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.158369] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52073a71-7914-c95d-531e-11974c394272/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1215.158657] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Uploaded image 088b6ede-5bb4-4485-8338-801656621e71 to the Glance image server {{(pid=65107) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1215.161192] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Destroying the VM {{(pid=65107) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1215.162349] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cba4b783-3054-4c04-9826-e80006ac8de4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.169961] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1215.169961] env[65107]: value = "task-5103791" [ 1215.169961] env[65107]: _type = "Task" [ 1215.169961] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.179266] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103791, 'name': Destroy_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.550337] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103790, 'name': ReconfigVM_Task, 'duration_secs': 0.507952} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.550693] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Reconfigured VM instance instance-00000079 to attach disk [datastore1] 2c80c51e-dff3-4e56-bc13-32b792f67d97/2c80c51e-dff3-4e56-bc13-32b792f67d97.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.553881] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b7be8df-d686-4acd-91a5-34d557502b95 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.562404] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1215.562404] env[65107]: value = "task-5103792" [ 1215.562404] env[65107]: _type = "Task" [ 1215.562404] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.573115] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103792, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.577717] env[65107]: DEBUG nova.compute.manager [req-418c53cc-5d5d-4651-9cdb-295e7b50619f req-590a8330-344d-4016-a8fe-f335e67ee282 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Received event network-vif-deleted-429e11fc-b198-49e7-bce8-d832c0a6d038 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1215.577717] env[65107]: INFO nova.compute.manager [req-418c53cc-5d5d-4651-9cdb-295e7b50619f req-590a8330-344d-4016-a8fe-f335e67ee282 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Neutron deleted interface 429e11fc-b198-49e7-bce8-d832c0a6d038; detaching it from the instance and deleting it from the info cache [ 1215.577985] env[65107]: DEBUG nova.network.neutron [req-418c53cc-5d5d-4651-9cdb-295e7b50619f req-590a8330-344d-4016-a8fe-f335e67ee282 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1215.601802] env[65107]: DEBUG nova.compute.manager [req-badc8e0c-a10a-4450-bbc3-d12fbd937561 req-f6034152-98a4-46cd-9aa8-c5d0ac602974 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-vif-plugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1215.601978] env[65107]: DEBUG oslo_concurrency.lockutils [req-badc8e0c-a10a-4450-bbc3-d12fbd937561 req-f6034152-98a4-46cd-9aa8-c5d0ac602974 service nova] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.602368] env[65107]: DEBUG oslo_concurrency.lockutils [req-badc8e0c-a10a-4450-bbc3-d12fbd937561 req-f6034152-98a4-46cd-9aa8-c5d0ac602974 service nova] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.602914] env[65107]: DEBUG oslo_concurrency.lockutils [req-badc8e0c-a10a-4450-bbc3-d12fbd937561 req-f6034152-98a4-46cd-9aa8-c5d0ac602974 service nova] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.603031] env[65107]: DEBUG nova.compute.manager [req-badc8e0c-a10a-4450-bbc3-d12fbd937561 req-f6034152-98a4-46cd-9aa8-c5d0ac602974 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] No waiting events found dispatching network-vif-plugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1215.603374] env[65107]: WARNING nova.compute.manager [req-badc8e0c-a10a-4450-bbc3-d12fbd937561 req-f6034152-98a4-46cd-9aa8-c5d0ac602974 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received unexpected event network-vif-plugged-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 for instance with vm_state shelved_offloaded and task_state spawning. [ 1215.649139] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.649139] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1215.649279] env[65107]: DEBUG nova.network.neutron [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1215.681684] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103791, 'name': Destroy_Task, 'duration_secs': 0.425671} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.681998] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Destroyed the VM [ 1215.682658] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deleting Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1215.682658] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8cae6eda-35ae-4808-b792-a7e0a1984a32 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.691944] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1215.691944] env[65107]: value = "task-5103793" [ 1215.691944] env[65107]: _type = "Task" [ 1215.691944] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.705858] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103793, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.716808] env[65107]: DEBUG nova.network.neutron [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1216.073323] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103792, 'name': Rename_Task, 'duration_secs': 0.171567} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.073692] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1216.073968] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7cbb0b3e-381f-4ff6-a127-5b3b664bb639 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.082780] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1216.082780] env[65107]: value = "task-5103794" [ 1216.082780] env[65107]: _type = "Task" [ 1216.082780] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.083034] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22b41a49-9116-48d0-a383-e80f45836d84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.094413] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103794, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.100615] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eceaf1c2-11b5-44ab-8f73-73fa1477e04f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.139430] env[65107]: DEBUG nova.compute.manager [req-418c53cc-5d5d-4651-9cdb-295e7b50619f req-590a8330-344d-4016-a8fe-f335e67ee282 service nova] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Detach interface failed, port_id=429e11fc-b198-49e7-bce8-d832c0a6d038, reason: Instance d7b4bc1e-c94e-4654-9345-2a8aa945a896 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1216.153996] env[65107]: WARNING neutronclient.v2_0.client [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1216.154728] env[65107]: WARNING openstack [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.155250] env[65107]: WARNING openstack [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.205413] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103793, 'name': RemoveSnapshot_Task, 'duration_secs': 0.36012} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.205747] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deleted Snapshot of the VM instance {{(pid=65107) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1216.206045] env[65107]: DEBUG nova.compute.manager [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1216.206898] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dd902e-ed61-41ad-9a7a-2414e541984b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.209913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.210146] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.210363] env[65107]: INFO nova.compute.manager [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Attaching volume c0acec4c-a8f0-4cbc-a146-0ce2a3362af3 to /dev/sdb [ 1216.218389] env[65107]: INFO nova.compute.manager [-] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Took 1.48 seconds to deallocate network for instance. [ 1216.247893] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfa1ba6-19df-4206-adf7-5cb004a05354 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.258131] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079e4faa-e294-448a-9d61-ef57cb344d8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.264824] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c74f3ec-9134-495f-a15e-621bb6353d43 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.294500] env[65107]: WARNING openstack [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.294657] env[65107]: WARNING openstack [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.303279] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e3f420-14ef-4646-b171-e9fac3c35af2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.309254] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301c6a1a-9c29-41d1-a9a8-2f01184d0078 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.316426] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d41baa5-8513-4891-89e4-eefb2d5136e5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.332293] env[65107]: DEBUG nova.compute.provider_tree [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.337572] env[65107]: DEBUG nova.virt.block_device [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating existing volume attachment record: eb3b8343-62fd-41ff-b907-b7365a0fe9ae {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1216.378436] env[65107]: WARNING neutronclient.v2_0.client [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1216.379300] env[65107]: WARNING openstack [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.380052] env[65107]: WARNING openstack [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.472579] env[65107]: DEBUG nova.network.neutron [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1216.597293] env[65107]: DEBUG oslo_vmware.api [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103794, 'name': PowerOnVM_Task, 'duration_secs': 0.489022} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.597613] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.597771] env[65107]: INFO nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Took 8.48 seconds to spawn the instance on the hypervisor. [ 1216.598311] env[65107]: DEBUG nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1216.598731] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04b314b-5c76-439d-95e2-e2a66fb35e73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.729981] env[65107]: INFO nova.compute.manager [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Shelve offloading [ 1216.732298] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.841323] env[65107]: DEBUG nova.scheduler.client.report [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1216.975261] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1217.005430] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='311d55bc651318a9ff10ef7bf5901293',container_format='bare',created_at=2025-12-12T13:31:02Z,direct_url=,disk_format='vmdk',id=e4b7eab7-f146-4e21-9038-cf84fc47c977,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1113377013-shelved',owner='f379144b78764fe394039d87b043a946',properties=ImageMetaProps,protected=,size=31664128,status='active',tags=,updated_at=2025-12-12T13:31:17Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1217.005752] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1217.005909] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1217.006104] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1217.006252] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1217.006398] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1217.006606] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1217.006761] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1217.006945] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1217.007099] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1217.007275] env[65107]: DEBUG nova.virt.hardware [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1217.008254] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301809c5-8860-4c70-93e5-dc5f1e83d181 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.016738] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10586bb0-2d07-4424-8935-b278342369c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.031119] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:b5:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd903c404-a23a-40c0-a217-96d4bb2e5b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b5b909b-b28a-42bf-ad8f-3dd0e7df6811', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1217.039296] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1217.039663] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1217.039922] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4f74279-5a15-4f0b-a689-88b18d4fa184 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.061180] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1217.061180] env[65107]: value = "task-5103798" [ 1217.061180] env[65107]: _type = "Task" [ 1217.061180] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.069736] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103798, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.115759] env[65107]: INFO nova.compute.manager [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Took 15.34 seconds to build instance. [ 1217.235898] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1217.236491] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a14b35a1-8568-42b7-92af-d72159cd7df5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.246237] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1217.246237] env[65107]: value = "task-5103799" [ 1217.246237] env[65107]: _type = "Task" [ 1217.246237] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.257463] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] VM already powered off {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1217.257747] env[65107]: DEBUG nova.compute.manager [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1217.258606] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd6941e-832d-4545-951a-537f32eadd8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.266659] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.266847] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1217.267137] env[65107]: DEBUG nova.network.neutron [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1217.347276] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.314s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.347587] env[65107]: WARNING neutronclient.v2_0.client [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1217.350367] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.695s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1217.350624] env[65107]: DEBUG nova.objects.instance [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lazy-loading 'resources' on Instance uuid 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.382486] env[65107]: INFO nova.network.neutron [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating port e48b19c2-6253-4a9c-a6fb-1daae3460ede with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1217.573329] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103798, 'name': CreateVM_Task, 'duration_secs': 0.371163} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.573439] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1217.573908] env[65107]: WARNING neutronclient.v2_0.client [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1217.574289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.574465] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1217.574835] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1217.575107] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38cbc2f9-c100-4558-ad77-278f916fcc9e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.582330] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1217.582330] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2498f-6719-6451-3b06-bce17b83f6f0" [ 1217.582330] env[65107]: _type = "Task" [ 1217.582330] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.589545] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2498f-6719-6451-3b06-bce17b83f6f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.617235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-52434ee9-28f6-4845-863e-f46ed94b1ef3 tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.848s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.689993] env[65107]: DEBUG nova.compute.manager [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1217.690165] env[65107]: DEBUG nova.compute.manager [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing instance network info cache due to event network-changed-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1217.690383] env[65107]: DEBUG oslo_concurrency.lockutils [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Acquiring lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.690525] env[65107]: DEBUG oslo_concurrency.lockutils [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Acquired lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1217.690685] env[65107]: DEBUG nova.network.neutron [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Refreshing network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1217.769986] env[65107]: WARNING neutronclient.v2_0.client [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1217.770844] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1217.770844] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1217.926566] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1217.927025] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.014884] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0847af7-dfcd-4054-bfc9-e14a58ce9092 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.023266] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646b1b77-dd18-43fe-982f-3c217d536e6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.054918] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdde45d2-f4d5-4e30-a102-2f03325756ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.058751] env[65107]: WARNING neutronclient.v2_0.client [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.059313] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.059660] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.073781] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ee9a44-2c58-4343-8142-af001715a0a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.087970] env[65107]: DEBUG nova.compute.provider_tree [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.098748] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1218.098995] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Processing image e4b7eab7-f146-4e21-9038-cf84fc47c977 {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1218.099242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.099384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.099562] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1218.100096] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e466d026-b4a5-4ee2-a13d-2f23f1a08b30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.111868] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1218.112066] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1218.112847] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32cb798-938b-4804-bd4f-1a90adb13ad3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.118896] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1218.118896] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcfe53-ec2c-dad5-d0b4-752f0bb49972" [ 1218.118896] env[65107]: _type = "Task" [ 1218.118896] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.128352] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52fcfe53-ec2c-dad5-d0b4-752f0bb49972, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.193672] env[65107]: WARNING neutronclient.v2_0.client [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.194460] env[65107]: WARNING openstack [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.194842] env[65107]: WARNING openstack [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.203613] env[65107]: DEBUG nova.network.neutron [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1218.326883] env[65107]: WARNING openstack [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.327382] env[65107]: WARNING openstack [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.391595] env[65107]: WARNING neutronclient.v2_0.client [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.392276] env[65107]: WARNING openstack [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.392667] env[65107]: WARNING openstack [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.422761] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.423030] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.423230] env[65107]: DEBUG nova.compute.manager [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1218.424150] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f49c921-b73e-4478-9ccc-f93602b572f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.431279] env[65107]: DEBUG nova.compute.manager [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1218.431839] env[65107]: DEBUG nova.objects.instance [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'flavor' on Instance uuid 2c80c51e-dff3-4e56-bc13-32b792f67d97 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.476107] env[65107]: DEBUG nova.network.neutron [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updated VIF entry in instance network info cache for port 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1218.476479] env[65107]: DEBUG nova.network.neutron [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [{"id": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "address": "fa:16:3e:06:b5:9e", "network": {"id": "65f204aa-768a-43b3-82dd-38a044e777ff", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1939340827-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f379144b78764fe394039d87b043a946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b5b909b-b2", "ovs_interfaceid": "4b5b909b-b28a-42bf-ad8f-3dd0e7df6811", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1218.593665] env[65107]: DEBUG nova.scheduler.client.report [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1218.629343] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1218.629633] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Fetch image to [datastore2] OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb/OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1218.629827] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Downloading stream optimized image e4b7eab7-f146-4e21-9038-cf84fc47c977 to [datastore2] OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb/OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb.vmdk on the data store datastore2 as vApp {{(pid=65107) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1218.629999] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Downloading image file data e4b7eab7-f146-4e21-9038-cf84fc47c977 to the ESX as VM named 'OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb' {{(pid=65107) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1218.707238] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1218.707658] env[65107]: WARNING neutronclient.v2_0.client [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.708330] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.708784] env[65107]: WARNING openstack [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.713871] env[65107]: WARNING neutronclient.v2_0.client [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.718605] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1218.718605] env[65107]: value = "resgroup-9" [ 1218.718605] env[65107]: _type = "ResourcePool" [ 1218.718605] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1218.718605] env[65107]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-88e65aa2-b4f3-4005-9521-b1d29efcd81e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.740752] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease: (returnval){ [ 1218.740752] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52346089-5886-5308-1230-d03d52e1b6fc" [ 1218.740752] env[65107]: _type = "HttpNfcLease" [ 1218.740752] env[65107]: } obtained for vApp import into resource pool (val){ [ 1218.740752] env[65107]: value = "resgroup-9" [ 1218.740752] env[65107]: _type = "ResourcePool" [ 1218.740752] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1218.741103] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the lease: (returnval){ [ 1218.741103] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52346089-5886-5308-1230-d03d52e1b6fc" [ 1218.741103] env[65107]: _type = "HttpNfcLease" [ 1218.741103] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1218.751218] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1218.751218] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52346089-5886-5308-1230-d03d52e1b6fc" [ 1218.751218] env[65107]: _type = "HttpNfcLease" [ 1218.751218] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1218.901546] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.901546] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.901546] env[65107]: DEBUG nova.network.neutron [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1218.950304] env[65107]: DEBUG nova.compute.manager [req-9dc40336-a45c-4f7f-a749-547ff75fb324 req-3ebb1e82-6ede-434c-aceb-10336aa6ca00 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-vif-plugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1218.950720] env[65107]: DEBUG oslo_concurrency.lockutils [req-9dc40336-a45c-4f7f-a749-547ff75fb324 req-3ebb1e82-6ede-434c-aceb-10336aa6ca00 service nova] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.950892] env[65107]: DEBUG oslo_concurrency.lockutils [req-9dc40336-a45c-4f7f-a749-547ff75fb324 req-3ebb1e82-6ede-434c-aceb-10336aa6ca00 service nova] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.951106] env[65107]: DEBUG oslo_concurrency.lockutils [req-9dc40336-a45c-4f7f-a749-547ff75fb324 req-3ebb1e82-6ede-434c-aceb-10336aa6ca00 service nova] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.951302] env[65107]: DEBUG nova.compute.manager [req-9dc40336-a45c-4f7f-a749-547ff75fb324 req-3ebb1e82-6ede-434c-aceb-10336aa6ca00 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] No waiting events found dispatching network-vif-plugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1218.951527] env[65107]: WARNING nova.compute.manager [req-9dc40336-a45c-4f7f-a749-547ff75fb324 req-3ebb1e82-6ede-434c-aceb-10336aa6ca00 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received unexpected event network-vif-plugged-e48b19c2-6253-4a9c-a6fb-1daae3460ede for instance with vm_state shelved_offloaded and task_state spawning. [ 1218.978842] env[65107]: DEBUG oslo_concurrency.lockutils [req-b10108cf-73e2-4b13-8167-0d42b342d51a req-b531b781-b897-4e1c-a4c4-87e3f0b22d86 service nova] Releasing lock "refresh_cache-93a1b356-b6d5-435b-b30f-ea5ade031b7e" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1219.098812] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.102041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.369s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.102041] env[65107]: DEBUG nova.objects.instance [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'resources' on Instance uuid d7b4bc1e-c94e-4654-9345-2a8aa945a896 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.121536] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1219.122407] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2e7187-4808-4135-88f5-2d903934afb5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.131778] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1219.132541] env[65107]: INFO nova.scheduler.client.report [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Deleted allocations for instance 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3 [ 1219.133607] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-183fdada-becc-431d-8e8e-f1f4c5d77357 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.219512] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1219.219765] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1219.219919] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleting the datastore file [datastore1] 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1219.220168] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e95ddfe3-902d-4546-87e1-27b2bce16fac {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.228381] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1219.228381] env[65107]: value = "task-5103803" [ 1219.228381] env[65107]: _type = "Task" [ 1219.228381] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.239558] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.248751] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1219.248751] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52346089-5886-5308-1230-d03d52e1b6fc" [ 1219.248751] env[65107]: _type = "HttpNfcLease" [ 1219.248751] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1219.249098] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1219.249098] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52346089-5886-5308-1230-d03d52e1b6fc" [ 1219.249098] env[65107]: _type = "HttpNfcLease" [ 1219.249098] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1219.249884] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cce3d03-9307-43e7-8e6c-a3af7894759c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.258270] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e4947-f481-9d2a-e254-0b8c15e8bf07/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1219.258460] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating HTTP connection to write to file with size = 31664128 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e4947-f481-9d2a-e254-0b8c15e8bf07/disk-0.vmdk. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1219.321070] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-263d4ffd-c193-4e45-bf3a-fc1b913a35c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.402711] env[65107]: WARNING neutronclient.v2_0.client [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.403495] env[65107]: WARNING openstack [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.403894] env[65107]: WARNING openstack [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.441170] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1219.441545] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b7c0273-cc4c-4696-971e-2845ef30c970 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.449801] env[65107]: DEBUG oslo_vmware.api [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1219.449801] env[65107]: value = "task-5103804" [ 1219.449801] env[65107]: _type = "Task" [ 1219.449801] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.463726] env[65107]: DEBUG oslo_vmware.api [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.534819] env[65107]: WARNING openstack [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.535292] env[65107]: WARNING openstack [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.614181] env[65107]: WARNING neutronclient.v2_0.client [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.615047] env[65107]: WARNING openstack [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.615259] env[65107]: WARNING openstack [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.645768] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f53c826f-9524-4390-9223-69485bda2e0e tempest-AttachVolumeNegativeTest-1325286790 tempest-AttachVolumeNegativeTest-1325286790-project-member] Lock "3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.629s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.716095] env[65107]: DEBUG nova.network.neutron [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1219.742949] env[65107]: DEBUG nova.compute.manager [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-vif-unplugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1219.742949] env[65107]: DEBUG oslo_concurrency.lockutils [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.743166] env[65107]: DEBUG oslo_concurrency.lockutils [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.743268] env[65107]: DEBUG oslo_concurrency.lockutils [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.743382] env[65107]: DEBUG nova.compute.manager [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] No waiting events found dispatching network-vif-unplugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1219.743747] env[65107]: WARNING nova.compute.manager [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received unexpected event network-vif-unplugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 for instance with vm_state shelved and task_state shelving_offloading. [ 1219.743747] env[65107]: DEBUG nova.compute.manager [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1219.743959] env[65107]: DEBUG nova.compute.manager [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing instance network info cache due to event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1219.744046] env[65107]: DEBUG oslo_concurrency.lockutils [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.744179] env[65107]: DEBUG oslo_concurrency.lockutils [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1219.744398] env[65107]: DEBUG nova.network.neutron [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1219.752047] env[65107]: DEBUG oslo_vmware.api [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135823} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.752735] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1219.753094] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1219.753390] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1219.775723] env[65107]: INFO nova.scheduler.client.report [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted allocations for instance 475851f6-24a2-4f0e-9567-d2a23947a5e3 [ 1219.794025] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742c87aa-964e-418f-9b2a-a109a6f36817 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.805438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b088500-c831-41e2-9e0a-160d85780800 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.843239] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bba498-20b5-4c26-8fc0-44c45b91bae8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.855381] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151d57e0-82a8-4e4d-ad6c-019b21254319 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.873204] env[65107]: DEBUG nova.compute.provider_tree [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.963052] env[65107]: DEBUG oslo_vmware.api [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103804, 'name': PowerOffVM_Task, 'duration_secs': 0.230299} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.965880] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1219.966106] env[65107]: DEBUG nova.compute.manager [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1219.966964] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a33917-9f39-4be2-acab-37282c2b823c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.223279] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1220.249752] env[65107]: WARNING neutronclient.v2_0.client [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.250738] env[65107]: WARNING openstack [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.251312] env[65107]: WARNING openstack [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.272132] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b7d904cccea73af94494926b2c0ebe3e',container_format='bare',created_at=2025-12-12T13:31:04Z,direct_url=,disk_format='vmdk',id=830c17b6-eb60-4df2-8862-7545a6ab062f,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1903848991-shelved',owner='782397ddd6cb40298dcf7f0da7353d67',properties=ImageMetaProps,protected=,size=31598080,status='active',tags=,updated_at=2025-12-12T13:31:19Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1220.272556] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1220.272773] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1220.273411] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1220.273661] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1220.273841] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1220.274307] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1220.274557] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1220.274741] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1220.275111] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1220.275111] env[65107]: DEBUG nova.virt.hardware [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1220.276389] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6559a97-96bc-4667-9e2e-959f38592a79 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.282203] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.289979] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e3f785-fc7f-460b-ad98-a9fbc0c98ad3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.308877] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:b0:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '17c839f5-4de0-449c-9a24-4e0e2fca37ca', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e48b19c2-6253-4a9c-a6fb-1daae3460ede', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1220.315908] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1220.318669] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1220.318924] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65865d0e-4afa-41ee-ab90-78aab3d8cf37 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.343504] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1220.343504] env[65107]: value = "task-5103805" [ 1220.343504] env[65107]: _type = "Task" [ 1220.343504] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.355893] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103805, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.377422] env[65107]: DEBUG nova.scheduler.client.report [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1220.481225] env[65107]: DEBUG oslo_concurrency.lockutils [None req-0a3d0b4d-27ca-4536-ac55-faf0cd5d07fb tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.058s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.488804] env[65107]: WARNING openstack [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.489280] env[65107]: WARNING openstack [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.569954] env[65107]: WARNING neutronclient.v2_0.client [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.570791] env[65107]: WARNING openstack [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.571266] env[65107]: WARNING openstack [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.671758] env[65107]: DEBUG nova.network.neutron [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updated VIF entry in instance network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1220.672153] env[65107]: DEBUG nova.network.neutron [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap72f482ac-0a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1220.685513] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1220.685992] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e4947-f481-9d2a-e254-0b8c15e8bf07/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1220.688187] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ac83e2-94e9-471d-86f3-9684e1cd2d42 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.695790] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e4947-f481-9d2a-e254-0b8c15e8bf07/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1220.696108] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e4947-f481-9d2a-e254-0b8c15e8bf07/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1220.696420] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-df6195c1-4ec6-45b5-929d-c70c9fb239c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.855806] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103805, 'name': CreateVM_Task, 'duration_secs': 0.380589} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.855990] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1220.856545] env[65107]: WARNING neutronclient.v2_0.client [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.856951] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.857148] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1220.857559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1220.857830] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-154c2e7b-56b0-4c17-9f8d-8d05c9078874 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.863546] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1220.863546] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0a397-72b3-1d2e-95d9-75a6619004aa" [ 1220.863546] env[65107]: _type = "Task" [ 1220.863546] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.872394] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f0a397-72b3-1d2e-95d9-75a6619004aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.883052] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.781s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.886696] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.604s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.886696] env[65107]: DEBUG nova.objects.instance [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'resources' on Instance uuid 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1220.910139] env[65107]: INFO nova.scheduler.client.report [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted allocations for instance d7b4bc1e-c94e-4654-9345-2a8aa945a896 [ 1220.916133] env[65107]: DEBUG oslo_vmware.rw_handles [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528e4947-f481-9d2a-e254-0b8c15e8bf07/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1220.916399] env[65107]: INFO nova.virt.vmwareapi.images [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Downloaded image file data e4b7eab7-f146-4e21-9038-cf84fc47c977 [ 1220.917327] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1691d501-f518-4f6a-a1bd-a3df82914e26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.934712] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10afc78e-3942-4ce1-8108-a12931e30e7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.974118] env[65107]: INFO nova.virt.vmwareapi.images [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] The imported VM was unregistered [ 1220.977049] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1220.977298] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1220.977583] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1dc2d6e7-bbe9-4b29-abe0-56ac5eb88b95 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.989886] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Created directory with path [datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1220.990111] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb/OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb.vmdk to [datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk. {{(pid=65107) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1220.990384] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-94234015-aa31-42eb-8964-1e4a582854c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.997331] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1220.997331] env[65107]: value = "task-5103808" [ 1220.997331] env[65107]: _type = "Task" [ 1220.997331] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.005966] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.009902] env[65107]: DEBUG nova.compute.manager [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-changed-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1221.009902] env[65107]: DEBUG nova.compute.manager [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Refreshing instance network info cache due to event network-changed-e48b19c2-6253-4a9c-a6fb-1daae3460ede. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1221.009902] env[65107]: DEBUG oslo_concurrency.lockutils [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.010047] env[65107]: DEBUG oslo_concurrency.lockutils [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1221.010167] env[65107]: DEBUG nova.network.neutron [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Refreshing network info cache for port e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1221.176827] env[65107]: DEBUG oslo_concurrency.lockutils [req-737625f1-a992-460d-90d9-b8aeeeb42e8d req-6a7eba27-5546-4bf3-9260-18a149c4df84 service nova] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1221.329469] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.329762] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1221.330437] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c80c51e-dff3-4e56-bc13-32b792f67d97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.330643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1221.330834] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.333187] env[65107]: INFO nova.compute.manager [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Terminating instance [ 1221.381029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1221.381029] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Processing image 830c17b6-eb60-4df2-8862-7545a6ab062f {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1221.381029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.381029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1221.381029] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1221.381029] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbfdd6cf-90a7-4302-9c8d-61b38e952269 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.392915] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1221.393650] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992903', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'name': 'volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '751ebb31-1a27-4630-9213-4312a6fa9298', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'serial': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1221.394274] env[65107]: DEBUG nova.objects.instance [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'numa_topology' on Instance uuid 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.396291] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694ec733-202c-4c10-ac87-76acefdd1f25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.401772] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1221.402062] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1221.403925] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6bdfdec-d6a7-480d-9739-3e02310d2463 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.424045] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eb4bd2-6fb3-48a7-bd10-0adc6897ede9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.427309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-1addffe1-0fb1-4c45-bc98-b892bd1abfc6 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "d7b4bc1e-c94e-4654-9345-2a8aa945a896" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.325s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.430620] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1221.430620] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52d28d77-0ac0-29ea-e371-7dcd558611be" [ 1221.430620] env[65107]: _type = "Task" [ 1221.430620] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.461346] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3/volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1221.462575] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4702ae6c-dd53-466b-95f6-cae63b67cedd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.482379] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1221.482823] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Fetch image to [datastore2] OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e/OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1221.483058] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Downloading stream optimized image 830c17b6-eb60-4df2-8862-7545a6ab062f to [datastore2] OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e/OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e.vmdk on the data store datastore2 as vApp {{(pid=65107) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1221.483238] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Downloading image file data 830c17b6-eb60-4df2-8862-7545a6ab062f to the ESX as VM named 'OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e' {{(pid=65107) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1221.489075] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1221.489075] env[65107]: value = "task-5103809" [ 1221.489075] env[65107]: _type = "Task" [ 1221.489075] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.500156] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103809, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.510175] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.515626] env[65107]: WARNING neutronclient.v2_0.client [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.516328] env[65107]: WARNING openstack [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.516652] env[65107]: WARNING openstack [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.588813] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1221.588813] env[65107]: value = "resgroup-9" [ 1221.588813] env[65107]: _type = "ResourcePool" [ 1221.588813] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1221.589146] env[65107]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-16373260-6c2c-44c8-bb7b-eed88104a88e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.616936] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lease: (returnval){ [ 1221.616936] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5764a-3ce6-b3af-a2ca-b29c95e7c687" [ 1221.616936] env[65107]: _type = "HttpNfcLease" [ 1221.616936] env[65107]: } obtained for vApp import into resource pool (val){ [ 1221.616936] env[65107]: value = "resgroup-9" [ 1221.616936] env[65107]: _type = "ResourcePool" [ 1221.616936] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1221.617353] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the lease: (returnval){ [ 1221.617353] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5764a-3ce6-b3af-a2ca-b29c95e7c687" [ 1221.617353] env[65107]: _type = "HttpNfcLease" [ 1221.617353] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1221.624258] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1221.624258] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5764a-3ce6-b3af-a2ca-b29c95e7c687" [ 1221.624258] env[65107]: _type = "HttpNfcLease" [ 1221.624258] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1221.630610] env[65107]: WARNING openstack [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.631033] env[65107]: WARNING openstack [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.701372] env[65107]: WARNING neutronclient.v2_0.client [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.702068] env[65107]: WARNING openstack [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.702466] env[65107]: WARNING openstack [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.797239] env[65107]: DEBUG nova.network.neutron [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updated VIF entry in instance network info cache for port e48b19c2-6253-4a9c-a6fb-1daae3460ede. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1221.797656] env[65107]: DEBUG nova.network.neutron [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1221.837676] env[65107]: DEBUG nova.compute.manager [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1221.837970] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1221.838978] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b5400f-2a02-463d-aa1d-7ef76c3f8527 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.849045] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1221.849438] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97920b95-207b-417b-bb91-eb99d226a98d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.903334] env[65107]: DEBUG nova.objects.base [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Object Instance<475851f6-24a2-4f0e-9567-d2a23947a5e3> lazy-loaded attributes: resources,numa_topology {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1221.933038] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1221.933038] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1221.933038] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore1] 2c80c51e-dff3-4e56-bc13-32b792f67d97 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1221.933038] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-023865bd-9a34-4b94-b271-2ce27448c4bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.940472] env[65107]: DEBUG oslo_vmware.api [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1221.940472] env[65107]: value = "task-5103812" [ 1221.940472] env[65107]: _type = "Task" [ 1221.940472] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.963748] env[65107]: DEBUG oslo_vmware.api [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.002673] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.012310] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.080508] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fa3da9-22f0-442a-9df3-aea121fc3a40 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.090308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af3b9fe-57c8-45a9-af42-0bdfb50a9bcf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.130399] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2b4f6c-ef8d-41fd-b0db-ddd1f8669981 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.140037] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1222.140037] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5764a-3ce6-b3af-a2ca-b29c95e7c687" [ 1222.140037] env[65107]: _type = "HttpNfcLease" [ 1222.140037] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1222.144693] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513dc932-3d44-425f-a492-1751d268e1ea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.168157] env[65107]: DEBUG nova.compute.provider_tree [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.301094] env[65107]: DEBUG oslo_concurrency.lockutils [req-36fc2920-bed7-4420-8633-df558dfce302 req-52a9b30f-7de7-44a5-a366-ba9d8b084a6e service nova] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1222.455246] env[65107]: DEBUG oslo_vmware.api [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.469054] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1222.504210] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.516614] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.543702] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1222.544088] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1222.635914] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1222.635914] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5764a-3ce6-b3af-a2ca-b29c95e7c687" [ 1222.635914] env[65107]: _type = "HttpNfcLease" [ 1222.635914] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1222.636319] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1222.636319] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f5764a-3ce6-b3af-a2ca-b29c95e7c687" [ 1222.636319] env[65107]: _type = "HttpNfcLease" [ 1222.636319] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1222.637043] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2444d8-acd3-48bf-b2c2-45bca8266525 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.647636] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525617c4-c6bc-1e6b-e992-7c28e533ae96/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1222.647822] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating HTTP connection to write to file with size = 31598080 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525617c4-c6bc-1e6b-e992-7c28e533ae96/disk-0.vmdk. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1222.710025] env[65107]: DEBUG nova.scheduler.client.report [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1222.725801] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-69ad9f72-778c-42a3-9383-e8b05fd8078d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.957734] env[65107]: DEBUG oslo_vmware.api [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.007483} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.958036] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1222.958309] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1222.958523] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1222.958702] env[65107]: INFO nova.compute.manager [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1222.958964] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1222.959199] env[65107]: DEBUG nova.compute.manager [-] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1222.959301] env[65107]: DEBUG nova.network.neutron [-] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1222.959565] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.960133] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.960391] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.008556] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.011248] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.028643] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.047623] env[65107]: INFO nova.compute.manager [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Detaching volume f451fb1b-2805-4784-8cf8-a2cce68d18d1 [ 1223.093222] env[65107]: INFO nova.virt.block_device [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Attempting to driver detach volume f451fb1b-2805-4784-8cf8-a2cce68d18d1 from mountpoint /dev/sdb [ 1223.093513] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1223.093721] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1223.097959] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144aa42e-006a-46ad-95b8-8feeacb15fe5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.130642] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200d45c0-57f6-4c05-9879-a014cac2a1ef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.144576] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e9faa8-e962-4584-89ad-06e5e976cfa8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.177710] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155f40e4-20a2-4187-becf-32f734841bec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.199842] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The volume has not been displaced from its original location: [datastore2] volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1/volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1223.206601] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1223.212469] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd9b6dab-269a-4576-8099-6039007cfb15 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.232636] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.346s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1223.245746] env[65107]: DEBUG oslo_vmware.api [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1223.245746] env[65107]: value = "task-5103813" [ 1223.245746] env[65107]: _type = "Task" [ 1223.245746] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.282588] env[65107]: DEBUG oslo_vmware.api [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103813, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.339037] env[65107]: DEBUG nova.compute.manager [req-680e7850-e769-4f04-8202-9252ffc4711a req-5b374a6e-b5a6-4977-8a01-4e37a763b7e6 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Received event network-vif-deleted-6af11f76-b4de-4137-b77f-a285924025d9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1223.339143] env[65107]: INFO nova.compute.manager [req-680e7850-e769-4f04-8202-9252ffc4711a req-5b374a6e-b5a6-4977-8a01-4e37a763b7e6 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Neutron deleted interface 6af11f76-b4de-4137-b77f-a285924025d9; detaching it from the instance and deleting it from the info cache [ 1223.339325] env[65107]: DEBUG nova.network.neutron [req-680e7850-e769-4f04-8202-9252ffc4711a req-5b374a6e-b5a6-4977-8a01-4e37a763b7e6 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.508228] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.525990] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.742727] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b8ad162d-ae7a-4449-8057-1e8731a8f5f7 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.448s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1223.743828] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.275s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1223.744052] env[65107]: INFO nova.compute.manager [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Unshelving [ 1223.757998] env[65107]: DEBUG oslo_vmware.api [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103813, 'name': ReconfigVM_Task, 'duration_secs': 0.31437} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.758333] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1223.763720] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-436ffe3b-8f42-46cd-a1ff-437a7d86ce26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.780020] env[65107]: DEBUG oslo_vmware.api [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1223.780020] env[65107]: value = "task-5103815" [ 1223.780020] env[65107]: _type = "Task" [ 1223.780020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.794150] env[65107]: DEBUG oslo_vmware.api [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103815, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.802294] env[65107]: DEBUG nova.network.neutron [-] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.842808] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34ad8ceb-b3bf-43d6-a703-ee050e30bd63 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.853339] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5349d12e-7d81-43b9-9956-f3a591012bd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.897458] env[65107]: DEBUG nova.compute.manager [req-680e7850-e769-4f04-8202-9252ffc4711a req-5b374a6e-b5a6-4977-8a01-4e37a763b7e6 service nova] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Detach interface failed, port_id=6af11f76-b4de-4137-b77f-a285924025d9, reason: Instance 2c80c51e-dff3-4e56-bc13-32b792f67d97 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1223.937356] env[65107]: DEBUG nova.compute.manager [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1224.003494] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103809, 'name': ReconfigVM_Task, 'duration_secs': 2.19427} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.005776] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Reconfigured VM instance instance-00000075 to attach disk [datastore2] volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3/volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1224.010841] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf415cf3-6fff-4e3d-9904-cbec853d3c1d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.022567] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1224.022784] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525617c4-c6bc-1e6b-e992-7c28e533ae96/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1224.026554] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b029469-fbdf-47b6-b3a1-dc7e66825242 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.035025] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525617c4-c6bc-1e6b-e992-7c28e533ae96/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1224.035207] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525617c4-c6bc-1e6b-e992-7c28e533ae96/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1224.038402] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-3ed1d5d5-9821-4299-9356-dab31819e657 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.039979] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103808, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.553428} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.041210] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb/OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb.vmdk to [datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk. [ 1224.041404] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Cleaning up location [datastore2] OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1224.041571] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_1b3abd90-238a-4558-b108-20d60a70b1cb {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1224.041875] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1224.041875] env[65107]: value = "task-5103816" [ 1224.041875] env[65107]: _type = "Task" [ 1224.041875] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.042427] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f082b6db-8191-4cbc-a007-8c20235b2d36 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.053294] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103816, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.054757] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1224.054757] env[65107]: value = "task-5103817" [ 1224.054757] env[65107]: _type = "Task" [ 1224.054757] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.063323] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.249805] env[65107]: DEBUG oslo_vmware.rw_handles [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525617c4-c6bc-1e6b-e992-7c28e533ae96/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1224.250129] env[65107]: INFO nova.virt.vmwareapi.images [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Downloaded image file data 830c17b6-eb60-4df2-8862-7545a6ab062f [ 1224.254311] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe7a9a2-e889-48cf-a261-18e1381eb4f3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.276562] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7eb09f30-66bd-416f-a7f9-52c700bdfdcd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.293745] env[65107]: DEBUG oslo_vmware.api [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103815, 'name': ReconfigVM_Task, 'duration_secs': 0.17002} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.294279] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992887', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'name': 'volume-f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bef91b92-add1-4df9-bc24-dab15ce04338', 'attached_at': '', 'detached_at': '', 'volume_id': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1', 'serial': 'f451fb1b-2805-4784-8cf8-a2cce68d18d1'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1224.304887] env[65107]: INFO nova.compute.manager [-] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Took 1.35 seconds to deallocate network for instance. [ 1224.308458] env[65107]: INFO nova.virt.vmwareapi.images [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] The imported VM was unregistered [ 1224.311360] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1224.311627] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Creating directory with path [datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1224.315574] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2051a5c-7213-4219-b1cf-882281966dc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.353200] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Created directory with path [datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1224.353415] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e/OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e.vmdk to [datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk. {{(pid=65107) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1224.353701] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-2823cddb-0d70-4dba-9dd7-4f750f553290 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.362600] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1224.362600] env[65107]: value = "task-5103819" [ 1224.362600] env[65107]: _type = "Task" [ 1224.362600] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.373604] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.462050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.462205] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.554331] env[65107]: DEBUG oslo_vmware.api [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103816, 'name': ReconfigVM_Task, 'duration_secs': 0.180958} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.554855] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992903', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'name': 'volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '751ebb31-1a27-4630-9213-4312a6fa9298', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'serial': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1224.564550] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.052988} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.564782] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1224.564942] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.565206] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk to [datastore2] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1224.565535] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4da2b5f-ad9c-486f-a388-572f7fbc4102 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.572595] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1224.572595] env[65107]: value = "task-5103820" [ 1224.572595] env[65107]: _type = "Task" [ 1224.572595] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.585516] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.773580] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.820802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.845861] env[65107]: DEBUG nova.objects.instance [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'flavor' on Instance uuid bef91b92-add1-4df9-bc24-dab15ce04338 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1224.874042] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.972633] env[65107]: INFO nova.compute.claims [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1225.085096] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.374478] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.480194] env[65107]: INFO nova.compute.resource_tracker [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating resource usage from migration b7c48e7e-98b8-47a3-83e5-8410160c5892 [ 1225.583796] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.637752] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e624d98b-3fb6-4844-83c2-73e6d85787b9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.645898] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2318bb41-6fa1-4786-9a05-6f63d4c7f8f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.650053] env[65107]: DEBUG nova.objects.instance [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.684195] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71851b21-3d54-4ade-b4ac-656f4a9e3ffd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.693340] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e7df8e-f51d-4b22-b78c-5227af5d48c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.710025] env[65107]: DEBUG nova.compute.provider_tree [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1225.855170] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5e8a68a6-3c69-44d9-a4e5-7a6f7d2a9ef3 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.311s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.875727] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.897017] env[65107]: DEBUG oslo_concurrency.lockutils [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1226.083470] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.156671] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d4549389-c913-4d4f-89b1-2124c841796b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.946s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.158214] env[65107]: DEBUG oslo_concurrency.lockutils [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.261s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.158546] env[65107]: DEBUG nova.compute.manager [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1226.160126] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96e894d-f2ce-413d-a471-46769fc795e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.169983] env[65107]: DEBUG nova.compute.manager [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1226.170853] env[65107]: DEBUG nova.objects.instance [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.213316] env[65107]: DEBUG nova.scheduler.client.report [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1226.375109] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.586128] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.720105] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.257s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.720105] env[65107]: INFO nova.compute.manager [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Migrating [ 1226.730530] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.957s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.730766] env[65107]: DEBUG nova.objects.instance [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'pci_requests' on Instance uuid 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.876485] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.919305] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1226.919735] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.919996] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "bef91b92-add1-4df9-bc24-dab15ce04338-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1226.920256] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.920469] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.922962] env[65107]: INFO nova.compute.manager [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Terminating instance [ 1227.094728] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.179906] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1227.180380] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24fd94b4-1cf2-4071-997d-37a2fe6b498d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.190088] env[65107]: DEBUG oslo_vmware.api [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1227.190088] env[65107]: value = "task-5103822" [ 1227.190088] env[65107]: _type = "Task" [ 1227.190088] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.201658] env[65107]: DEBUG oslo_vmware.api [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.235176] env[65107]: DEBUG nova.objects.instance [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'numa_topology' on Instance uuid 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1227.236845] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1227.237058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1227.237250] env[65107]: DEBUG nova.network.neutron [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1227.376184] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.427943] env[65107]: DEBUG nova.compute.manager [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1227.428168] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1227.429132] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa625924-89dc-48cc-9c33-d228d279b94b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.437882] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1227.438378] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55601ea3-244e-498e-9ea9-b2137626751b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.446016] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1227.446016] env[65107]: value = "task-5103823" [ 1227.446016] env[65107]: _type = "Task" [ 1227.446016] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.455217] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103823, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.589891] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.701931] env[65107]: DEBUG oslo_vmware.api [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.739688] env[65107]: INFO nova.compute.claims [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1227.742614] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1227.743432] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.743782] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.861813] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.862310] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.878908] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.936917] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1227.937610] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.937964] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.956736] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103823, 'name': PowerOffVM_Task, 'duration_secs': 0.229031} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.957029] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1227.957208] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1227.957483] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45660e58-5feb-48b1-881a-6f016fa81d75 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.029892] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1228.030241] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1228.030241] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore2] bef91b92-add1-4df9-bc24-dab15ce04338 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.030510] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c115fab7-9cca-4856-a221-61a3c7da9e2f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.038625] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1228.038625] env[65107]: value = "task-5103825" [ 1228.038625] env[65107]: _type = "Task" [ 1228.038625] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.041599] env[65107]: DEBUG nova.network.neutron [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1228.055921] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.089360] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.205074] env[65107]: DEBUG oslo_vmware.api [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103822, 'name': PowerOffVM_Task, 'duration_secs': 0.952993} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.205587] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1228.205921] env[65107]: DEBUG nova.compute.manager [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1228.207196] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f3da3b-f1b4-458e-bb2c-fdf135814e7f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.380418] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.549335] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1228.558685] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.593017] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.723141] env[65107]: DEBUG oslo_concurrency.lockutils [None req-28d64af4-1c38-441e-a949-5b62ad6ffc2a tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.564s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1228.883340] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103819, 'name': MoveVirtualDisk_Task, 'duration_secs': 4.394196} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.883629] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e/OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e.vmdk to [datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk. [ 1228.883820] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Cleaning up location [datastore2] OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1228.883981] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_5ff4638b-2adc-482e-a200-08a06583a19e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.884289] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2bb91e9-b025-40a6-aa55-224049752e3a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.892290] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1228.892290] env[65107]: value = "task-5103826" [ 1228.892290] env[65107]: _type = "Task" [ 1228.892290] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.902359] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.920400] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845013c7-d526-4ed8-b0f1-ec1f838a3c28 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.929060] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39094ea5-fd02-4fb9-93e4-bfcd3bbefbd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.968871] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f10e051-57e0-43f4-b69a-9bd35de42337 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.980949] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33307576-a8c4-44d3-b918-dfa051b851b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.998404] env[65107]: DEBUG nova.compute.provider_tree [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1229.057706] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103825, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.090688] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.404132] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332446} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.404470] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1229.404683] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "[datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1229.404918] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk to [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1229.405237] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e27725ca-c783-4a0d-a744-37a8f66b38e4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.412421] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1229.412421] env[65107]: value = "task-5103827" [ 1229.412421] env[65107]: _type = "Task" [ 1229.412421] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.420983] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103827, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.478015] env[65107]: DEBUG nova.objects.instance [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.503059] env[65107]: DEBUG nova.scheduler.client.report [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1229.554235] env[65107]: DEBUG oslo_vmware.api [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103825, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.277129} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.554662] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1229.554891] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1229.555094] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1229.555274] env[65107]: INFO nova.compute.manager [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Took 2.13 seconds to destroy the instance on the hypervisor. [ 1229.555556] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1229.555765] env[65107]: DEBUG nova.compute.manager [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1229.555900] env[65107]: DEBUG nova.network.neutron [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1229.556208] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1229.556993] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1229.557351] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1229.592049] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103820, 'name': CopyVirtualDisk_Task, 'duration_secs': 4.698046} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.592378] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e4b7eab7-f146-4e21-9038-cf84fc47c977/e4b7eab7-f146-4e21-9038-cf84fc47c977.vmdk to [datastore2] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1229.593300] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abae344e-7cc3-4a50-be0f-035ece675e09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.597292] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1229.623231] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1229.623970] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f024691-176b-4d89-b913-fd7edea2f2bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.645403] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1229.645403] env[65107]: value = "task-5103828" [ 1229.645403] env[65107]: _type = "Task" [ 1229.645403] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.657164] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103828, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.923631] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103827, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.984242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.984622] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1229.984896] env[65107]: DEBUG nova.network.neutron [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1229.985145] env[65107]: DEBUG nova.objects.instance [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'info_cache' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1230.007897] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.277s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.008356] env[65107]: WARNING neutronclient.v2_0.client [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.011120] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.190s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.011436] env[65107]: DEBUG nova.objects.instance [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'resources' on Instance uuid 2c80c51e-dff3-4e56-bc13-32b792f67d97 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1230.045857] env[65107]: INFO nova.network.neutron [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1230.076978] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f01af3-3213-4783-885e-8bbf13526023 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.085691] env[65107]: DEBUG nova.compute.manager [req-198d4a5d-9361-4a9e-8851-b42f5307b01c req-0303f945-3f30-4595-bfb0-7a145e2d13af service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Received event network-vif-deleted-40c68565-653b-4c27-83db-52fff0de54f0 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1230.085691] env[65107]: INFO nova.compute.manager [req-198d4a5d-9361-4a9e-8851-b42f5307b01c req-0303f945-3f30-4595-bfb0-7a145e2d13af service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Neutron deleted interface 40c68565-653b-4c27-83db-52fff0de54f0; detaching it from the instance and deleting it from the info cache [ 1230.085691] env[65107]: DEBUG nova.network.neutron [req-198d4a5d-9361-4a9e-8851-b42f5307b01c req-0303f945-3f30-4595-bfb0-7a145e2d13af service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1230.106719] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1230.156401] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103828, 'name': ReconfigVM_Task, 'duration_secs': 0.390275} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.156607] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 93a1b356-b6d5-435b-b30f-ea5ade031b7e/93a1b356-b6d5-435b-b30f-ea5ade031b7e.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1230.158000] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3decdcf9-24af-4ae7-a293-4b71ad299020 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.167035] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1230.167035] env[65107]: value = "task-5103829" [ 1230.167035] env[65107]: _type = "Task" [ 1230.167035] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.177048] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103829, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.426157] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103827, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.488883] env[65107]: DEBUG nova.objects.base [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Object Instance<751ebb31-1a27-4630-9213-4312a6fa9298> lazy-loaded attributes: flavor,info_cache {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1230.553188] env[65107]: DEBUG nova.network.neutron [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1230.589029] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a439a8ec-ec0f-4dce-b665-2645df1e16c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.607496] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f660944-99d1-4db1-ada1-5406d6df05c1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.619999] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1230.620921] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12e76538-fec5-4994-884d-22482ad06119 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.631048] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1230.631048] env[65107]: value = "task-5103830" [ 1230.631048] env[65107]: _type = "Task" [ 1230.631048] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.651682] env[65107]: DEBUG nova.compute.manager [req-198d4a5d-9361-4a9e-8851-b42f5307b01c req-0303f945-3f30-4595-bfb0-7a145e2d13af service nova] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Detach interface failed, port_id=40c68565-653b-4c27-83db-52fff0de54f0, reason: Instance bef91b92-add1-4df9-bc24-dab15ce04338 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1230.658582] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.681372] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103829, 'name': Rename_Task, 'duration_secs': 0.178009} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.681853] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1230.682286] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0088295d-8b29-4bf3-ba68-22e2b2cc3494 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.693536] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1230.693536] env[65107]: value = "task-5103831" [ 1230.693536] env[65107]: _type = "Task" [ 1230.693536] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.707047] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.708567] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bd6fbe-f078-45c8-a3cc-26839462dfe2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.719320] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aaab1b3-6413-430a-b7f7-16bce69d00a5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.758612] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e911ea-4f0b-49eb-8c45-8a99d84164ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.769367] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72df1124-09d0-4f1e-9a70-42e3845791d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.792319] env[65107]: DEBUG nova.compute.provider_tree [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.932063] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103827, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.991701] env[65107]: WARNING neutronclient.v2_0.client [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.992693] env[65107]: WARNING openstack [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1230.993157] env[65107]: WARNING openstack [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1231.055603] env[65107]: INFO nova.compute.manager [-] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Took 1.50 seconds to deallocate network for instance. [ 1231.146461] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103830, 'name': PowerOffVM_Task, 'duration_secs': 0.231012} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.146869] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1231.147266] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1231.206557] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103831, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.296142] env[65107]: DEBUG nova.scheduler.client.report [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1231.428499] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103827, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.563796] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1231.647093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.647093] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1231.647636] env[65107]: DEBUG nova.network.neutron [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1231.656205] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1231.656205] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1231.656582] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1231.657867] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1231.657867] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1231.657867] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1231.658030] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1231.658228] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1231.658554] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1231.658701] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1231.658986] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1231.668221] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-798b27a9-fdcb-4905-89f8-dd25615380bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.688909] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1231.688909] env[65107]: value = "task-5103832" [ 1231.688909] env[65107]: _type = "Task" [ 1231.688909] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.707354] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103832, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.715178] env[65107]: DEBUG oslo_vmware.api [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103831, 'name': PowerOnVM_Task, 'duration_secs': 0.52909} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.715820] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1231.802490] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.808378] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.242s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.808378] env[65107]: DEBUG nova.objects.instance [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'resources' on Instance uuid bef91b92-add1-4df9-bc24-dab15ce04338 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1231.832271] env[65107]: WARNING openstack [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1231.832842] env[65107]: WARNING openstack [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1231.842180] env[65107]: INFO nova.scheduler.client.report [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance 2c80c51e-dff3-4e56-bc13-32b792f67d97 [ 1231.868680] env[65107]: DEBUG nova.compute.manager [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1231.869728] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67577e1-9cfc-4e4e-bceb-f5d08cbafca9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.927536] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103827, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.406018} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.932540] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/830c17b6-eb60-4df2-8862-7545a6ab062f/830c17b6-eb60-4df2-8862-7545a6ab062f.vmdk to [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1231.933843] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef8a245-3929-4e91-8b18-d2327f10c220 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.969715] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1231.969893] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1240aacc-6c6f-45c7-bc46-017951f84d6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.988062] env[65107]: WARNING neutronclient.v2_0.client [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1231.988821] env[65107]: WARNING openstack [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1231.989217] env[65107]: WARNING openstack [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.000178] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1232.000178] env[65107]: value = "task-5103833" [ 1232.000178] env[65107]: _type = "Task" [ 1232.000178] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.012443] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103833, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.085977] env[65107]: DEBUG nova.network.neutron [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [{"id": "13b14810-c694-4e47-b090-8c3bf39ad510", "address": "fa:16:3e:e7:23:0e", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13b14810-c6", "ovs_interfaceid": "13b14810-c694-4e47-b090-8c3bf39ad510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1232.132766] env[65107]: DEBUG nova.compute.manager [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-vif-plugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1232.132916] env[65107]: DEBUG oslo_concurrency.lockutils [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1232.133150] env[65107]: DEBUG oslo_concurrency.lockutils [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1232.133343] env[65107]: DEBUG oslo_concurrency.lockutils [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.133501] env[65107]: DEBUG nova.compute.manager [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] No waiting events found dispatching network-vif-plugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1232.133668] env[65107]: WARNING nova.compute.manager [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received unexpected event network-vif-plugged-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 for instance with vm_state shelved_offloaded and task_state spawning. [ 1232.133828] env[65107]: DEBUG nova.compute.manager [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1232.133980] env[65107]: DEBUG nova.compute.manager [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing instance network info cache due to event network-changed-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1232.134164] env[65107]: DEBUG oslo_concurrency.lockutils [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Acquiring lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1232.152618] env[65107]: WARNING neutronclient.v2_0.client [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1232.155674] env[65107]: WARNING openstack [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.155674] env[65107]: WARNING openstack [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.199119] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103832, 'name': ReconfigVM_Task, 'duration_secs': 0.3785} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.201931] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1232.313544] env[65107]: WARNING openstack [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.313974] env[65107]: WARNING openstack [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.361035] env[65107]: DEBUG oslo_concurrency.lockutils [None req-01ce275f-768e-4460-8345-2f0de2134c1c tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c80c51e-dff3-4e56-bc13-32b792f67d97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.031s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.389575] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b70add2c-9bc7-41c5-ae08-c7feea5bc021 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.415s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.413581] env[65107]: WARNING neutronclient.v2_0.client [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1232.414595] env[65107]: WARNING openstack [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1232.414701] env[65107]: WARNING openstack [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1232.479095] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764bcfbe-49f7-44da-ba39-c40e87901d6f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.487716] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7702d98-9b12-4e2b-bcf5-25951cd3d177 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.527122] env[65107]: DEBUG nova.network.neutron [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1232.533050] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8720d014-bb0e-4574-af12-144825494edf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.546260] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3eb424-27d7-4c4f-88b1-c02f3b9718c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.550604] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103833, 'name': ReconfigVM_Task, 'duration_secs': 0.549605} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.551210] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Reconfigured VM instance instance-00000071 to attach disk [datastore2] bcced66b-1464-4834-892b-3d0e652da31a/bcced66b-1464-4834-892b-3d0e652da31a.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1232.552180] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55421654-691a-4fd7-9a77-8057cb5e3c72 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.563294] env[65107]: DEBUG nova.compute.provider_tree [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1232.565922] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1232.565922] env[65107]: value = "task-5103834" [ 1232.565922] env[65107]: _type = "Task" [ 1232.565922] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.574702] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103834, 'name': Rename_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.588810] env[65107]: DEBUG oslo_concurrency.lockutils [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1232.711065] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1232.711065] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1232.711541] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1232.711541] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1232.711541] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1232.711774] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1232.711869] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1232.712022] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1232.712189] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1232.712372] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1232.712555] env[65107]: DEBUG nova.virt.hardware [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1232.717826] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Reconfiguring VM instance instance-00000031 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1232.718137] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb976975-45d9-44b0-a6b3-67b7ad34690f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.737735] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1232.737735] env[65107]: value = "task-5103835" [ 1232.737735] env[65107]: _type = "Task" [ 1232.737735] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.746196] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.036482] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1233.039025] env[65107]: DEBUG oslo_concurrency.lockutils [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Acquired lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1233.039138] env[65107]: DEBUG nova.network.neutron [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Refreshing network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1233.064687] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='44cfe20e12a809562533fb504c7006c5',container_format='bare',created_at=2025-12-12T13:31:17Z,direct_url=,disk_format='vmdk',id=088b6ede-5bb4-4485-8338-801656621e71,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1870429067-shelved',owner='dced4f9aef2f49cf990203b693533aa1',properties=ImageMetaProps,protected=,size=31673856,status='active',tags=,updated_at=2025-12-12T13:31:32Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1233.064980] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1233.065155] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1233.065341] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1233.065523] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1233.065684] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1233.065891] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1233.066062] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1233.066237] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1233.066421] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1233.066570] env[65107]: DEBUG nova.virt.hardware [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1233.067753] env[65107]: DEBUG nova.scheduler.client.report [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1233.071906] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66dfbd4-4a34-4382-abe8-68c81232c839 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.083806] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103834, 'name': Rename_Task, 'duration_secs': 0.137231} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.086664] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1233.087656] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8475735-f53c-4141-95b1-24de840d04f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.090113] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fca3460-ac76-4d1a-8976-e0a896e110e1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.108626] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:c2:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72f482ac-0a93-4cf4-9d54-a7e812ca4cd9', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1233.115874] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1233.117787] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1233.118148] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1233.118148] env[65107]: value = "task-5103836" [ 1233.118148] env[65107]: _type = "Task" [ 1233.118148] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.118354] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f160191-5da1-40cd-ae20-ca6e093eca1a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.145377] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103836, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.147276] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1233.147276] env[65107]: value = "task-5103837" [ 1233.147276] env[65107]: _type = "Task" [ 1233.147276] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.156921] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103837, 'name': CreateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.251640] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103835, 'name': ReconfigVM_Task, 'duration_secs': 0.201435} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.252101] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Reconfigured VM instance instance-00000031 to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1233.252937] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b72eff-9f53-4bee-99aa-469437a8963f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.277882] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a/cd481deb-b12c-47ab-9ab1-0c4bdbb2024a.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1233.278332] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f98fead6-df3b-4a30-989f-3b3503456d30 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.298786] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1233.298786] env[65107]: value = "task-5103838" [ 1233.298786] env[65107]: _type = "Task" [ 1233.298786] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.308611] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103838, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.541886] env[65107]: WARNING neutronclient.v2_0.client [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1233.543046] env[65107]: WARNING openstack [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1233.543398] env[65107]: WARNING openstack [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1233.572225] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.597805] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1233.598012] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc4969fa-c40b-491f-a711-8c678f744161 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.600622] env[65107]: INFO nova.scheduler.client.report [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted allocations for instance bef91b92-add1-4df9-bc24-dab15ce04338 [ 1233.607829] env[65107]: DEBUG oslo_vmware.api [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1233.607829] env[65107]: value = "task-5103839" [ 1233.607829] env[65107]: _type = "Task" [ 1233.607829] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.616826] env[65107]: DEBUG oslo_vmware.api [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103839, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.643382] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103836, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.660487] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103837, 'name': CreateVM_Task} progress is 25%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.662274] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.662593] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.662913] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.663250] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.663781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.666535] env[65107]: INFO nova.compute.manager [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Terminating instance [ 1233.678851] env[65107]: WARNING openstack [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1233.679654] env[65107]: WARNING openstack [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1233.764772] env[65107]: WARNING neutronclient.v2_0.client [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1233.765335] env[65107]: WARNING openstack [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1233.765744] env[65107]: WARNING openstack [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1233.815560] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103838, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.858502] env[65107]: DEBUG nova.network.neutron [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updated VIF entry in instance network info cache for port 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1233.858883] env[65107]: DEBUG nova.network.neutron [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [{"id": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "address": "fa:16:3e:de:c2:03", "network": {"id": "c74bc521-ccb3-4cc2-adfe-a4b28b1ab0da", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2064598886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dced4f9aef2f49cf990203b693533aa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72f482ac-0a", "ovs_interfaceid": "72f482ac-0a93-4cf4-9d54-a7e812ca4cd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1233.953802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.954132] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.954371] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.955163] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.955384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.957837] env[65107]: INFO nova.compute.manager [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Terminating instance [ 1234.108414] env[65107]: DEBUG oslo_concurrency.lockutils [None req-be374a88-a100-4b81-ae5c-6e9e4803980a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "bef91b92-add1-4df9-bc24-dab15ce04338" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.189s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.120377] env[65107]: DEBUG oslo_vmware.api [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103839, 'name': PowerOnVM_Task, 'duration_secs': 0.421856} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.120648] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1234.120880] env[65107]: DEBUG nova.compute.manager [None req-10573e80-d6d3-4ee0-b877-660f23076128 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1234.121795] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c90b55-bd63-43e6-91fb-82290c397a6d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.144360] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103836, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.159033] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103837, 'name': CreateVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.173570] env[65107]: DEBUG nova.compute.manager [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1234.173860] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1234.175581] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfeaa1c-0f25-43c3-90fb-d709a1cc50c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.185063] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1234.185358] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10093f68-497f-4e01-ba9d-cea4db374b6d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.194335] env[65107]: DEBUG oslo_vmware.api [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1234.194335] env[65107]: value = "task-5103840" [ 1234.194335] env[65107]: _type = "Task" [ 1234.194335] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.204377] env[65107]: DEBUG oslo_vmware.api [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103840, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.313052] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103838, 'name': ReconfigVM_Task, 'duration_secs': 0.922733} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.313438] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Reconfigured VM instance instance-00000031 to attach disk [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a/cd481deb-b12c-47ab-9ab1-0c4bdbb2024a.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1234.313752] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1234.361610] env[65107]: DEBUG oslo_concurrency.lockutils [req-90a35a29-f33c-45e7-92d7-4a2683d8333a req-c6d33267-35f2-4264-a48e-3a3de8a6c83d service nova] Releasing lock "refresh_cache-475851f6-24a2-4f0e-9567-d2a23947a5e3" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1234.462524] env[65107]: DEBUG nova.compute.manager [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1234.462753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1234.463718] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed47d1a7-a9d4-4c0a-ac1f-5cb88521d960 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.472526] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1234.472815] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c5ac28c-3b1f-4056-b12a-78ff87e6a527 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.480331] env[65107]: DEBUG oslo_vmware.api [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1234.480331] env[65107]: value = "task-5103841" [ 1234.480331] env[65107]: _type = "Task" [ 1234.480331] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.490760] env[65107]: DEBUG oslo_vmware.api [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103841, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.649975] env[65107]: DEBUG oslo_vmware.api [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103836, 'name': PowerOnVM_Task, 'duration_secs': 1.288411} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.653492] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1234.663056] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103837, 'name': CreateVM_Task, 'duration_secs': 1.057639} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.663977] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1234.664674] env[65107]: WARNING neutronclient.v2_0.client [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1234.665151] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.665373] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1234.665802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1234.666561] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d99d9f85-59e5-42d9-9327-31c7e5fdd689 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.673213] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1234.673213] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527fc231-867a-32e3-45c6-f35e26f321b2" [ 1234.673213] env[65107]: _type = "Task" [ 1234.673213] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.683695] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527fc231-867a-32e3-45c6-f35e26f321b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.707194] env[65107]: DEBUG oslo_vmware.api [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103840, 'name': PowerOffVM_Task, 'duration_secs': 0.258083} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.707573] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1234.707775] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1234.708026] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be1a02b6-d82d-4094-8471-9973b22b13ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.759611] env[65107]: DEBUG nova.compute.manager [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1234.760558] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13b04ca-d356-484d-b7f3-f5fbcc64318e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.780374] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1234.780711] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1234.780870] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleting the datastore file [datastore2] 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1234.781218] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3385a3ec-f2f4-400b-9fad-e3b9a3fa0e8e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.791123] env[65107]: DEBUG oslo_vmware.api [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for the task: (returnval){ [ 1234.791123] env[65107]: value = "task-5103843" [ 1234.791123] env[65107]: _type = "Task" [ 1234.791123] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.800581] env[65107]: DEBUG oslo_vmware.api [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.823743] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ac56f7-d741-410f-b03e-b87f70773321 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.844685] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27badd8e-e25c-4d4d-89c8-f98681e1e998 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.863879] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1234.991493] env[65107]: DEBUG oslo_vmware.api [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103841, 'name': PowerOffVM_Task, 'duration_secs': 0.201909} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.991493] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1234.991778] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1234.991935] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ff93a69-f2b0-455e-b241-2589c756ad07 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.057672] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1235.057891] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1235.058086] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleting the datastore file [datastore1] 2c45d232-a948-4ad7-80a0-589c2dc91ec1 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.058364] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cede8267-29b4-4089-b134-248f5bcafc73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.065275] env[65107]: DEBUG oslo_vmware.api [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for the task: (returnval){ [ 1235.065275] env[65107]: value = "task-5103845" [ 1235.065275] env[65107]: _type = "Task" [ 1235.065275] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.074197] env[65107]: DEBUG oslo_vmware.api [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103845, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.183691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1235.183970] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Processing image 088b6ede-5bb4-4485-8338-801656621e71 {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1235.184225] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.184372] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquired lock "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1235.184651] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1235.184929] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bc7ec23-8121-41b9-82f8-70b9220a691c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.194377] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1235.194612] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1235.195459] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61e5c112-6bd0-454c-b5c3-cf24409d5ed4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.201644] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1235.201644] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c8f6c-b8de-e0b1-7c0c-a8cad8051f21" [ 1235.201644] env[65107]: _type = "Task" [ 1235.201644] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.210224] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]520c8f6c-b8de-e0b1-7c0c-a8cad8051f21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.280137] env[65107]: DEBUG oslo_concurrency.lockutils [None req-4e014fe2-4ec9-447a-bc9f-602578e0c181 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.017s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.301333] env[65107]: DEBUG oslo_vmware.api [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Task: {'id': task-5103843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360843} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.302319] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.302552] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1235.302839] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1235.302973] env[65107]: INFO nova.compute.manager [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1235.303295] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1235.303816] env[65107]: DEBUG nova.compute.manager [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1235.303996] env[65107]: DEBUG nova.network.neutron [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1235.304360] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.305017] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1235.305335] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.370587] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.370955] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.374588] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.407112] env[65107]: DEBUG nova.network.neutron [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Port cef8867b-7dd6-425a-b5db-760494cf6eb4 binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1235.576395] env[65107]: DEBUG oslo_vmware.api [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Task: {'id': task-5103845, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251018} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.576671] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.576989] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1235.577260] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1235.577462] env[65107]: INFO nova.compute.manager [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1235.577715] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1235.577920] env[65107]: DEBUG nova.compute.manager [-] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1235.578035] env[65107]: DEBUG nova.network.neutron [-] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1235.578288] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.578814] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1235.579090] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.628247] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.714539] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Preparing fetch location {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1235.714840] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Fetch image to [datastore2] OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c/OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c.vmdk {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1235.715037] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Downloading stream optimized image 088b6ede-5bb4-4485-8338-801656621e71 to [datastore2] OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c/OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c.vmdk on the data store datastore2 as vApp {{(pid=65107) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1235.715215] env[65107]: DEBUG nova.virt.vmwareapi.images [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Downloading image file data 088b6ede-5bb4-4485-8338-801656621e71 to the ESX as VM named 'OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c' {{(pid=65107) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1235.722273] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.722822] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.804044] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1235.804044] env[65107]: value = "resgroup-9" [ 1235.804044] env[65107]: _type = "ResourcePool" [ 1235.804044] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1235.804977] env[65107]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-382a3880-3717-48d8-b572-afbb1601913b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.828242] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease: (returnval){ [ 1235.828242] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1235.828242] env[65107]: _type = "HttpNfcLease" [ 1235.828242] env[65107]: } obtained for vApp import into resource pool (val){ [ 1235.828242] env[65107]: value = "resgroup-9" [ 1235.828242] env[65107]: _type = "ResourcePool" [ 1235.828242] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1235.828906] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the lease: (returnval){ [ 1235.828906] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1235.828906] env[65107]: _type = "HttpNfcLease" [ 1235.828906] env[65107]: } to be ready. {{(pid=65107) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1235.836076] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1235.836076] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1235.836076] env[65107]: _type = "HttpNfcLease" [ 1235.836076] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1235.930825] env[65107]: DEBUG nova.compute.manager [req-3bdb0620-86cf-47d3-981f-d4a7181f8688 req-94d64b0c-8ae5-4b8e-9a3f-463260a9ec5d service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Received event network-vif-deleted-4b5b909b-b28a-42bf-ad8f-3dd0e7df6811 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1235.930825] env[65107]: INFO nova.compute.manager [req-3bdb0620-86cf-47d3-981f-d4a7181f8688 req-94d64b0c-8ae5-4b8e-9a3f-463260a9ec5d service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Neutron deleted interface 4b5b909b-b28a-42bf-ad8f-3dd0e7df6811; detaching it from the instance and deleting it from the info cache [ 1235.930825] env[65107]: DEBUG nova.network.neutron [req-3bdb0620-86cf-47d3-981f-d4a7181f8688 req-94d64b0c-8ae5-4b8e-9a3f-463260a9ec5d service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1235.953682] env[65107]: DEBUG nova.compute.manager [req-d02d0b65-cd94-4035-a927-dd8425009e16 req-5dd71847-2883-408c-905e-15a1a399d50b service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Received event network-vif-deleted-966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1235.953885] env[65107]: INFO nova.compute.manager [req-d02d0b65-cd94-4035-a927-dd8425009e16 req-5dd71847-2883-408c-905e-15a1a399d50b service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Neutron deleted interface 966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87; detaching it from the instance and deleting it from the info cache [ 1235.954063] env[65107]: DEBUG nova.network.neutron [req-d02d0b65-cd94-4035-a927-dd8425009e16 req-5dd71847-2883-408c-905e-15a1a399d50b service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.225327] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1236.338146] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1236.338146] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1236.338146] env[65107]: _type = "HttpNfcLease" [ 1236.338146] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1236.377396] env[65107]: DEBUG nova.network.neutron [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.380425] env[65107]: DEBUG nova.network.neutron [-] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.430682] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.430988] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.431191] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.436744] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd61e010-c061-4903-ab37-a682c7e0b568 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.448485] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48233f2f-e2db-4135-b8f1-ebf54067a57c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.461032] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6aee6121-2bb1-4cd7-9d7e-26c5673f8b16 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.472416] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fda38b4-239e-458b-97e4-e1cdebd0e9c4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.494518] env[65107]: DEBUG nova.compute.manager [req-3bdb0620-86cf-47d3-981f-d4a7181f8688 req-94d64b0c-8ae5-4b8e-9a3f-463260a9ec5d service nova] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Detach interface failed, port_id=4b5b909b-b28a-42bf-ad8f-3dd0e7df6811, reason: Instance 93a1b356-b6d5-435b-b30f-ea5ade031b7e could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1236.516082] env[65107]: DEBUG nova.compute.manager [req-d02d0b65-cd94-4035-a927-dd8425009e16 req-5dd71847-2883-408c-905e-15a1a399d50b service nova] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Detach interface failed, port_id=966ebe8f-6f1a-49a5-b3bf-1a1bf51f8c87, reason: Instance 2c45d232-a948-4ad7-80a0-589c2dc91ec1 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1236.750029] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.750326] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.752225] env[65107]: INFO nova.compute.claims [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1236.838401] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1236.838401] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1236.838401] env[65107]: _type = "HttpNfcLease" [ 1236.838401] env[65107]: } is initializing. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1236.880950] env[65107]: INFO nova.compute.manager [-] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Took 1.58 seconds to deallocate network for instance. [ 1236.883913] env[65107]: INFO nova.compute.manager [-] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Took 1.31 seconds to deallocate network for instance. [ 1236.899054] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493fb145-76b8-4bf5-ac18-32943046d489 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.907569] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Suspending the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1236.907858] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d9954b6e-064f-4091-9bd2-457e518bcd85 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.915735] env[65107]: DEBUG oslo_vmware.api [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1236.915735] env[65107]: value = "task-5103847" [ 1236.915735] env[65107]: _type = "Task" [ 1236.915735] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.930531] env[65107]: DEBUG oslo_vmware.api [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103847, 'name': SuspendVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.338983] env[65107]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1237.338983] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1237.338983] env[65107]: _type = "HttpNfcLease" [ 1237.338983] env[65107]: } is ready. {{(pid=65107) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1237.339364] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1237.339364] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2eb57-10ce-bc1c-0556-ffbd56fb1bf0" [ 1237.339364] env[65107]: _type = "HttpNfcLease" [ 1237.339364] env[65107]: }. {{(pid=65107) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1237.340026] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4348f621-038d-4606-a866-5fd6c78e1992 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.348469] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e011f8-d3f2-2d0c-dd06-4f411b63a734/disk-0.vmdk from lease info. {{(pid=65107) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1237.348672] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating HTTP connection to write to file with size = 31673856 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e011f8-d3f2-2d0c-dd06-4f411b63a734/disk-0.vmdk. {{(pid=65107) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1237.405770] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1237.406910] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1237.412241] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-90821b52-d8b5-4b75-8ed9-1fc468bebdd3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.429811] env[65107]: DEBUG oslo_vmware.api [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103847, 'name': SuspendVM_Task} progress is 58%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.441369] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1237.484252] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.484582] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1237.484845] env[65107]: DEBUG nova.network.neutron [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1237.909130] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6efb593-a0d5-40ab-bc76-e4d25ea756cb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.917892] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c170af-85ab-45cf-8d45-599282f0108a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.930909] env[65107]: DEBUG oslo_vmware.api [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103847, 'name': SuspendVM_Task, 'duration_secs': 0.671995} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.956491] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Suspended the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1237.956701] env[65107]: DEBUG nova.compute.manager [None req-738e7e76-90b7-4c60-b8ef-592bc96965c5 tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1237.962919] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8104e69-a015-45b3-ab2e-f2ae0b436e97 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.966319] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3491e1-6db2-441a-bcae-f862e6dee65a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.978705] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d756edb7-4261-46c3-aa7e-e6f351ba12a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.988396] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1237.989143] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1237.989642] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.012921] env[65107]: DEBUG nova.compute.provider_tree [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1238.123460] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.123938] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.191535] env[65107]: WARNING neutronclient.v2_0.client [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1238.192254] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.192632] env[65107]: WARNING openstack [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.295051] env[65107]: DEBUG nova.network.neutron [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1238.518129] env[65107]: DEBUG nova.scheduler.client.report [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1238.588722] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Completed reading data from the image iterator. {{(pid=65107) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1238.588965] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e011f8-d3f2-2d0c-dd06-4f411b63a734/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1238.590027] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a86a1c-a28e-4550-ad82-28ed9028df79 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.597477] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e011f8-d3f2-2d0c-dd06-4f411b63a734/disk-0.vmdk is in state: ready. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1238.597668] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e011f8-d3f2-2d0c-dd06-4f411b63a734/disk-0.vmdk. {{(pid=65107) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1238.597924] env[65107]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-854add19-43ea-407f-8372-bb2557fcac03 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.798353] env[65107]: DEBUG oslo_concurrency.lockutils [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1239.024514] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1239.025105] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1239.027944] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.622s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1239.028198] env[65107]: DEBUG nova.objects.instance [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lazy-loading 'resources' on Instance uuid 93a1b356-b6d5-435b-b30f-ea5ade031b7e {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1239.296628] env[65107]: DEBUG oslo_vmware.rw_handles [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e011f8-d3f2-2d0c-dd06-4f411b63a734/disk-0.vmdk. {{(pid=65107) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1239.296870] env[65107]: INFO nova.virt.vmwareapi.images [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Downloaded image file data 088b6ede-5bb4-4485-8338-801656621e71 [ 1239.297751] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457ef81a-3997-4b41-a307-0226e6c17688 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.316659] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11fb94e9-8dbd-4d6e-b7ce-b163b98c779d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.335292] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb0ddb4-0718-497f-9f63-9c48748c636e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.357214] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bfd8e2-9f67-44ce-b32d-a34db81b4126 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.365842] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1239.369705] env[65107]: INFO nova.compute.manager [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Resuming [ 1239.370519] env[65107]: DEBUG nova.objects.instance [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'flavor' on Instance uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1239.428664] env[65107]: INFO nova.virt.vmwareapi.images [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] The imported VM was unregistered [ 1239.431465] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Caching image {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1239.431729] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Creating directory with path [datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1239.432995] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-090184da-fd6e-4fc9-b41e-fe766b874b73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.461110] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Created directory with path [datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71 {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1239.461345] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c/OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c.vmdk to [datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk. {{(pid=65107) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1239.461631] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-50093800-5342-4058-bc9e-798b9ca3cd51 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.469924] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1239.469924] env[65107]: value = "task-5103849" [ 1239.469924] env[65107]: _type = "Task" [ 1239.469924] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.479816] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.531267] env[65107]: DEBUG nova.compute.utils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1239.533251] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1239.533524] env[65107]: DEBUG nova.network.neutron [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1239.534050] env[65107]: WARNING neutronclient.v2_0.client [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.534542] env[65107]: WARNING neutronclient.v2_0.client [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.535615] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.536237] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.547227] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1239.598609] env[65107]: DEBUG nova.policy [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d55cff40f12420b8320448083ca8f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7784762e36ae4c6fb30012ec801c3b88', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1239.792546] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8d891f-750c-4ac5-889d-8fbbfcf580d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.805691] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9da476-fae9-468d-a563-ef500cf681c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.842619] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ae0057-ce49-4ab6-b868-51dc82456e26 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.851813] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fbe66a-c3a5-4706-ac09-908b060a61ce {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.869924] env[65107]: DEBUG nova.compute.provider_tree [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.874903] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1239.879377] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-116c6cd7-f5c7-4505-bbcc-b7f0eb229f88 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.890275] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1239.890275] env[65107]: value = "task-5103850" [ 1239.890275] env[65107]: _type = "Task" [ 1239.890275] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.900621] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.930520] env[65107]: DEBUG nova.network.neutron [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Successfully created port: d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1239.981980] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.373172] env[65107]: DEBUG nova.scheduler.client.report [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1240.408694] env[65107]: DEBUG oslo_vmware.api [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103850, 'name': PowerOnVM_Task, 'duration_secs': 0.508545} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.409137] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1240.409407] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-79ed5ca6-ad7f-4b1e-a33a-979d7919b0b3 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance 'cd481deb-b12c-47ab-9ab1-0c4bdbb2024a' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1240.485425] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.560558] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1240.591065] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1240.591519] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1240.591826] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1240.592179] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1240.592609] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1240.592900] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1240.593325] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.593621] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1240.593944] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1240.594269] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1240.594654] env[65107]: DEBUG nova.virt.hardware [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1240.595727] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65367075-9337-4435-a68e-b9cddca4bd0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.606137] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bffb58-eb68-46e1-8b68-e38001f5c77d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.880040] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.852s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.883529] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.477s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.883948] env[65107]: DEBUG nova.objects.instance [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lazy-loading 'resources' on Instance uuid 2c45d232-a948-4ad7-80a0-589c2dc91ec1 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.885853] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.886022] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquired lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1240.886207] env[65107]: DEBUG nova.network.neutron [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1240.907948] env[65107]: INFO nova.scheduler.client.report [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Deleted allocations for instance 93a1b356-b6d5-435b-b30f-ea5ade031b7e [ 1240.985474] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.390142] env[65107]: WARNING neutronclient.v2_0.client [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.390992] env[65107]: WARNING openstack [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.391441] env[65107]: WARNING openstack [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.416205] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f4cfb28e-88f5-4a8a-939b-3de4a6ea6687 tempest-ServerActionsTestOtherB-1046242315 tempest-ServerActionsTestOtherB-1046242315-project-member] Lock "93a1b356-b6d5-435b-b30f-ea5ade031b7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.754s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1241.486984] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.511784] env[65107]: DEBUG nova.compute.manager [req-6d1ec884-7197-4b08-8522-73026e447e00 req-b1e8fbbe-6548-408c-bdf2-edf5f0be7dd3 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Received event network-vif-plugged-d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1241.512053] env[65107]: DEBUG oslo_concurrency.lockutils [req-6d1ec884-7197-4b08-8522-73026e447e00 req-b1e8fbbe-6548-408c-bdf2-edf5f0be7dd3 service nova] Acquiring lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1241.512278] env[65107]: DEBUG oslo_concurrency.lockutils [req-6d1ec884-7197-4b08-8522-73026e447e00 req-b1e8fbbe-6548-408c-bdf2-edf5f0be7dd3 service nova] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1241.512582] env[65107]: DEBUG oslo_concurrency.lockutils [req-6d1ec884-7197-4b08-8522-73026e447e00 req-b1e8fbbe-6548-408c-bdf2-edf5f0be7dd3 service nova] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1241.512739] env[65107]: DEBUG nova.compute.manager [req-6d1ec884-7197-4b08-8522-73026e447e00 req-b1e8fbbe-6548-408c-bdf2-edf5f0be7dd3 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] No waiting events found dispatching network-vif-plugged-d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1241.512980] env[65107]: WARNING nova.compute.manager [req-6d1ec884-7197-4b08-8522-73026e447e00 req-b1e8fbbe-6548-408c-bdf2-edf5f0be7dd3 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Received unexpected event network-vif-plugged-d4ef2ece-8832-4198-a869-6869223cf87a for instance with vm_state building and task_state spawning. [ 1241.545052] env[65107]: WARNING openstack [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.545387] env[65107]: WARNING openstack [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.558277] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0e499d-df4f-45bc-bb88-e1d251199f2e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.571424] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de743d50-9261-4aa4-8c60-8618e4841623 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.611707] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530fc951-74df-4c45-be31-66927c305347 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.623196] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fa8b41-6237-48da-a70b-5897f89e1c39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.637595] env[65107]: DEBUG nova.network.neutron [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Successfully updated port: d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1241.647624] env[65107]: DEBUG nova.compute.provider_tree [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1241.651031] env[65107]: WARNING neutronclient.v2_0.client [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.651031] env[65107]: WARNING openstack [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.651690] env[65107]: WARNING openstack [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.661876] env[65107]: DEBUG nova.scheduler.client.report [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1241.746210] env[65107]: DEBUG nova.network.neutron [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [{"id": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "address": "fa:16:3e:1e:b0:b4", "network": {"id": "5153ef70-5eb6-49d2-a4b8-44eef67962e7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1363236739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "782397ddd6cb40298dcf7f0da7353d67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "17c839f5-4de0-449c-9a24-4e0e2fca37ca", "external-id": "nsx-vlan-transportzone-198", "segmentation_id": 198, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48b19c2-62", "ovs_interfaceid": "e48b19c2-6253-4a9c-a6fb-1daae3460ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1241.991602] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.150562] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.150747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1242.150925] env[65107]: DEBUG nova.network.neutron [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1242.168260] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.284s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.193611] env[65107]: INFO nova.scheduler.client.report [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Deleted allocations for instance 2c45d232-a948-4ad7-80a0-589c2dc91ec1 [ 1242.249481] env[65107]: DEBUG oslo_concurrency.lockutils [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Releasing lock "refresh_cache-bcced66b-1464-4834-892b-3d0e652da31a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.251308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e52d97-8b33-439b-b15b-8e7cd24028d4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.265219] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Resuming the VM {{(pid=65107) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1242.266225] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1aa21123-2e50-4e78-a568-3425d86a8aca {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.283600] env[65107]: DEBUG oslo_vmware.api [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1242.283600] env[65107]: value = "task-5103852" [ 1242.283600] env[65107]: _type = "Task" [ 1242.283600] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.298308] env[65107]: DEBUG oslo_vmware.api [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.483198] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103849, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.996163} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.483577] env[65107]: INFO nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c/OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c.vmdk to [datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk. [ 1242.483748] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Cleaning up location [datastore2] OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1242.483932] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_34bde554-6bf3-45e3-ab92-420444b9970c {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1242.484229] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de276794-64b0-49f1-99a9-3d7c1b52ba33 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.491795] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1242.491795] env[65107]: value = "task-5103853" [ 1242.491795] env[65107]: _type = "Task" [ 1242.491795] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.501353] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103853, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.640101] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.640401] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.640619] env[65107]: DEBUG nova.compute.manager [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Going to confirm migration 8 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1242.653864] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.654496] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.696473] env[65107]: DEBUG nova.network.neutron [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1242.703704] env[65107]: DEBUG oslo_concurrency.lockutils [None req-39646651-c6e0-46e4-8bce-734cca67e86a tempest-ServersTestJSON-191426844 tempest-ServersTestJSON-191426844-project-member] Lock "2c45d232-a948-4ad7-80a0-589c2dc91ec1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.749s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.721648] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.722112] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.796397] env[65107]: DEBUG oslo_vmware.api [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103852, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.803397] env[65107]: WARNING neutronclient.v2_0.client [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1242.804138] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.804536] env[65107]: WARNING openstack [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.938203] env[65107]: DEBUG nova.network.neutron [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updating instance_info_cache with network_info: [{"id": "d4ef2ece-8832-4198-a869-6869223cf87a", "address": "fa:16:3e:e0:61:86", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ef2ece-88", "ovs_interfaceid": "d4ef2ece-8832-4198-a869-6869223cf87a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1243.003936] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103853, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039894} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.004196] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1243.004379] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Releasing lock "[datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1243.004690] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk to [datastore2] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1243.004987] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-332b394a-3e9b-41e2-bdcd-3902c529043e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.014512] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1243.014512] env[65107]: value = "task-5103854" [ 1243.014512] env[65107]: _type = "Task" [ 1243.014512] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.024632] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.146987] env[65107]: WARNING neutronclient.v2_0.client [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.217280] env[65107]: WARNING neutronclient.v2_0.client [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.217643] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.217799] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.217971] env[65107]: DEBUG nova.network.neutron [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1243.218175] env[65107]: DEBUG nova.objects.instance [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'info_cache' on Instance uuid cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.297597] env[65107]: DEBUG oslo_vmware.api [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103852, 'name': PowerOnVM_Task, 'duration_secs': 0.714954} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.297919] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Resumed the VM {{(pid=65107) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1243.298117] env[65107]: DEBUG nova.compute.manager [None req-6676cb22-c292-424a-90e7-78919dbf52fc tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1243.298964] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0c5f9c-4e34-4d69-a537-785f1767a310 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.445027] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1243.445027] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Instance network_info: |[{"id": "d4ef2ece-8832-4198-a869-6869223cf87a", "address": "fa:16:3e:e0:61:86", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ef2ece-88", "ovs_interfaceid": "d4ef2ece-8832-4198-a869-6869223cf87a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1243.445027] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:61:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4ef2ece-8832-4198-a869-6869223cf87a', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1243.452245] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1243.452534] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1243.452802] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ab4b822-c31e-427e-9072-39e211ce1891 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.484685] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1243.484685] env[65107]: value = "task-5103855" [ 1243.484685] env[65107]: _type = "Task" [ 1243.484685] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.501334] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103855, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.526498] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.552029] env[65107]: DEBUG nova.compute.manager [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Received event network-changed-d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1243.552029] env[65107]: DEBUG nova.compute.manager [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Refreshing instance network info cache due to event network-changed-d4ef2ece-8832-4198-a869-6869223cf87a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1243.552186] env[65107]: DEBUG oslo_concurrency.lockutils [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Acquiring lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.552337] env[65107]: DEBUG oslo_concurrency.lockutils [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Acquired lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.552538] env[65107]: DEBUG nova.network.neutron [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Refreshing network info cache for port d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1243.997337] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103855, 'name': CreateVM_Task, 'duration_secs': 0.440457} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.997559] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1243.998089] env[65107]: WARNING neutronclient.v2_0.client [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.998442] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.998599] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.998931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1243.999216] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3da4ab1-5a7f-497d-9257-11b83a8aefeb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.007242] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1244.007242] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f2fea-0183-37c3-1cf3-8c5f8b5e27db" [ 1244.007242] env[65107]: _type = "Task" [ 1244.007242] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.020880] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f2fea-0183-37c3-1cf3-8c5f8b5e27db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.031865] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.056743] env[65107]: WARNING neutronclient.v2_0.client [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.057428] env[65107]: WARNING openstack [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.057777] env[65107]: WARNING openstack [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.193322] env[65107]: WARNING openstack [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.193907] env[65107]: WARNING openstack [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.228376] env[65107]: WARNING neutronclient.v2_0.client [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.228376] env[65107]: WARNING openstack [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.228376] env[65107]: WARNING openstack [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.280431] env[65107]: WARNING neutronclient.v2_0.client [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.281099] env[65107]: WARNING openstack [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.281432] env[65107]: WARNING openstack [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.453744] env[65107]: DEBUG nova.network.neutron [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updated VIF entry in instance network info cache for port d4ef2ece-8832-4198-a869-6869223cf87a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1244.454900] env[65107]: DEBUG nova.network.neutron [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updating instance_info_cache with network_info: [{"id": "d4ef2ece-8832-4198-a869-6869223cf87a", "address": "fa:16:3e:e0:61:86", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ef2ece-88", "ovs_interfaceid": "d4ef2ece-8832-4198-a869-6869223cf87a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1244.486279] env[65107]: WARNING openstack [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.486734] env[65107]: WARNING openstack [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.521654] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]525f2fea-0183-37c3-1cf3-8c5f8b5e27db, 'name': SearchDatastore_Task, 'duration_secs': 0.04902} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.525655] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.525970] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1244.526250] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.526399] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1244.526618] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1244.526946] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fa51c18-4eb8-44ae-957c-dc4146916318 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.537168] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.546492] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1244.546492] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1244.546492] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c40b3cc8-fad4-4308-8a57-3ffef9903e92 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.558231] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1244.558231] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2c38a-9fd6-3c97-a27b-df5d90fa9a43" [ 1244.558231] env[65107]: _type = "Task" [ 1244.558231] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.569852] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2c38a-9fd6-3c97-a27b-df5d90fa9a43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.585891] env[65107]: WARNING neutronclient.v2_0.client [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.586691] env[65107]: WARNING openstack [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.587084] env[65107]: WARNING openstack [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.693931] env[65107]: DEBUG nova.network.neutron [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [{"id": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "address": "fa:16:3e:59:3d:af", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcef8867b-7d", "ovs_interfaceid": "cef8867b-7dd6-425a-b5db-760494cf6eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1244.958086] env[65107]: DEBUG oslo_concurrency.lockutils [req-62440bb5-91f4-4a52-9a15-fb6b3a9fa333 req-86ca5b03-515d-445f-963f-b220d5821636 service nova] Releasing lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1245.031629] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.069883] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52e2c38a-9fd6-3c97-a27b-df5d90fa9a43, 'name': SearchDatastore_Task, 'duration_secs': 0.050701} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.070861] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8802391-32de-4f9b-9a41-0e3ae0fc009a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.078781] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1245.078781] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52973697-b474-7a7b-82d4-7e52d79e8704" [ 1245.078781] env[65107]: _type = "Task" [ 1245.078781] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.089372] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52973697-b474-7a7b-82d4-7e52d79e8704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.198461] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1245.198752] env[65107]: DEBUG nova.objects.instance [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'migration_context' on Instance uuid cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.534783] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.592223] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52973697-b474-7a7b-82d4-7e52d79e8704, 'name': SearchDatastore_Task, 'duration_secs': 0.049053} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.592693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1245.593125] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e72d0cb3-027c-49af-aa23-b791ae2f5fcf/e72d0cb3-027c-49af-aa23-b791ae2f5fcf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1245.593846] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ed9f0be-b718-4dba-898b-310526b1aa3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.603598] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1245.603598] env[65107]: value = "task-5103856" [ 1245.603598] env[65107]: _type = "Task" [ 1245.603598] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.616022] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103856, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.702110] env[65107]: DEBUG nova.objects.base [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1245.703235] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be92cfa-2df6-4171-8fe1-64ec51c9911f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.723561] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa1bca4e-d4e8-44e1-b420-3c98a4c2ab8b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.730431] env[65107]: DEBUG oslo_vmware.api [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1245.730431] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5240f352-27ba-34e0-dd13-1fb7078a8051" [ 1245.730431] env[65107]: _type = "Task" [ 1245.730431] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.740073] env[65107]: DEBUG oslo_vmware.api [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5240f352-27ba-34e0-dd13-1fb7078a8051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.032921] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103854, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.764488} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.033253] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/088b6ede-5bb4-4485-8338-801656621e71/088b6ede-5bb4-4485-8338-801656621e71.vmdk to [datastore2] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1246.034133] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694cc7ff-3785-4103-b2b6-9809c55c0d69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.057499] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.057839] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46c99c49-43e6-4143-beba-42569efbd133 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.078849] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1246.078849] env[65107]: value = "task-5103857" [ 1246.078849] env[65107]: _type = "Task" [ 1246.078849] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.091747] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103857, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.114615] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103856, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.241672] env[65107]: DEBUG oslo_vmware.api [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5240f352-27ba-34e0-dd13-1fb7078a8051, 'name': SearchDatastore_Task, 'duration_secs': 0.034347} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.242016] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.242289] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.589963] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103857, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.614872] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103856, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591506} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.615229] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] e72d0cb3-027c-49af-aa23-b791ae2f5fcf/e72d0cb3-027c-49af-aa23-b791ae2f5fcf.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1246.615508] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1246.615787] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8cfe07df-030c-4df1-93a1-38a7d2683429 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.624477] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1246.624477] env[65107]: value = "task-5103858" [ 1246.624477] env[65107]: _type = "Task" [ 1246.624477] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.636886] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.853114] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedfe249-6a08-4d67-a415-5799fcf0096e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.861136] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5063812a-58e8-48ea-93d0-0b482794afb1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.894145] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4a9e16-bc9a-4fa8-a354-c2e9ba2e8965 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.901894] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e535e77e-85d5-4729-8487-b0725dfdb2e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.916695] env[65107]: DEBUG nova.compute.provider_tree [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1247.093501] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103857, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.136274] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074289} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.136274] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1247.137228] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c049d6-9c79-4cba-acc7-922a1ee9dcc0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.160398] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] e72d0cb3-027c-49af-aa23-b791ae2f5fcf/e72d0cb3-027c-49af-aa23-b791ae2f5fcf.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1247.160736] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56d4d915-fe8b-49b2-8485-639d1526df3e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.181949] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1247.181949] env[65107]: value = "task-5103859" [ 1247.181949] env[65107]: _type = "Task" [ 1247.181949] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.190784] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.421034] env[65107]: DEBUG nova.scheduler.client.report [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1247.591446] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103857, 'name': ReconfigVM_Task, 'duration_secs': 1.031122} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.591794] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 475851f6-24a2-4f0e-9567-d2a23947a5e3/475851f6-24a2-4f0e-9567-d2a23947a5e3.vmdk or device None with type streamOptimized {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.592455] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93dec57b-32cb-467d-8797-45a803febe91 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.599066] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1247.599066] env[65107]: value = "task-5103860" [ 1247.599066] env[65107]: _type = "Task" [ 1247.599066] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.607798] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103860, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.692607] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103859, 'name': ReconfigVM_Task, 'duration_secs': 0.324262} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.692997] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Reconfigured VM instance instance-0000007a to attach disk [datastore1] e72d0cb3-027c-49af-aa23-b791ae2f5fcf/e72d0cb3-027c-49af-aa23-b791ae2f5fcf.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.693957] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-524cdd5a-8a09-4ddd-824a-bc6d27729520 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.701437] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1247.701437] env[65107]: value = "task-5103861" [ 1247.701437] env[65107]: _type = "Task" [ 1247.701437] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.710061] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103861, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.109691] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103860, 'name': Rename_Task, 'duration_secs': 0.395809} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.110224] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1248.110703] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86fce976-6827-409f-b927-cd7c7b806efd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.117175] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1248.117175] env[65107]: value = "task-5103862" [ 1248.117175] env[65107]: _type = "Task" [ 1248.117175] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.125719] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103862, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.211530] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103861, 'name': Rename_Task, 'duration_secs': 0.151085} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.211757] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1248.212022] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d15327f-7bef-4e88-a161-7faad0d0b9fe {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.218566] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1248.218566] env[65107]: value = "task-5103863" [ 1248.218566] env[65107]: _type = "Task" [ 1248.218566] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.228354] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.353232] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.353623] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.353891] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "bcced66b-1464-4834-892b-3d0e652da31a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.354104] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.354293] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.356789] env[65107]: INFO nova.compute.manager [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Terminating instance [ 1248.432131] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.190s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.629575] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103862, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.729983] env[65107]: DEBUG oslo_vmware.api [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103863, 'name': PowerOnVM_Task, 'duration_secs': 0.485262} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.729983] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1248.729983] env[65107]: INFO nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Took 8.17 seconds to spawn the instance on the hypervisor. [ 1248.729983] env[65107]: DEBUG nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1248.730597] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a12176e-9604-403e-b00e-0096884e1911 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.860892] env[65107]: DEBUG nova.compute.manager [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1248.861131] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1248.862057] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f227e127-0d0e-4b72-b7bd-86f23fa9bb0f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.870779] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1248.871078] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-416cabfc-909d-468f-b3c5-8c6f10f737aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.878646] env[65107]: DEBUG oslo_vmware.api [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1248.878646] env[65107]: value = "task-5103864" [ 1248.878646] env[65107]: _type = "Task" [ 1248.878646] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.889098] env[65107]: DEBUG oslo_vmware.api [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.998728] env[65107]: INFO nova.scheduler.client.report [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted allocation for migration b7c48e7e-98b8-47a3-83e5-8410160c5892 [ 1249.128555] env[65107]: DEBUG oslo_vmware.api [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103862, 'name': PowerOnVM_Task, 'duration_secs': 0.602792} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.129212] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1249.237442] env[65107]: DEBUG nova.compute.manager [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1249.238489] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c724efa4-b3b6-4738-88d6-97cf54c66266 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.252539] env[65107]: INFO nova.compute.manager [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Took 12.52 seconds to build instance. [ 1249.389884] env[65107]: DEBUG oslo_vmware.api [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103864, 'name': PowerOffVM_Task, 'duration_secs': 0.434308} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.390095] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1249.390312] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1249.390529] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d7179cd-375f-44c5-a5f0-270636041365 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.478438] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1249.478764] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1249.479129] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleting the datastore file [datastore2] bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1249.479438] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5fbc9294-0699-4ac3-97e3-a9f5f7792b7e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.489334] env[65107]: DEBUG oslo_vmware.api [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for the task: (returnval){ [ 1249.489334] env[65107]: value = "task-5103866" [ 1249.489334] env[65107]: _type = "Task" [ 1249.489334] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.498657] env[65107]: DEBUG oslo_vmware.api [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.504889] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d309ac88-3355-4d8e-acd2-4c8ae1f9fb72 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.864s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.758928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d934f55a-bf73-46eb-ba01-38e68275b87f tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.036s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.763256] env[65107]: DEBUG oslo_concurrency.lockutils [None req-62c0fc75-43c4-4e4f-b82d-7965e578b1cd tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 26.019s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.849929] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1249.850238] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1249.850457] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1249.850640] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1249.850805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.856610] env[65107]: INFO nova.compute.manager [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Terminating instance [ 1250.000947] env[65107]: DEBUG oslo_vmware.api [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Task: {'id': task-5103866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398316} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.001282] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1250.001617] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1250.001714] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1250.001834] env[65107]: INFO nova.compute.manager [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1250.002099] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1250.002300] env[65107]: DEBUG nova.compute.manager [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1250.002401] env[65107]: DEBUG nova.network.neutron [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1250.002681] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.003234] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1250.003490] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1250.065288] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.083757] env[65107]: DEBUG nova.compute.manager [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Received event network-changed-d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1250.083917] env[65107]: DEBUG nova.compute.manager [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Refreshing instance network info cache due to event network-changed-d4ef2ece-8832-4198-a869-6869223cf87a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1250.084147] env[65107]: DEBUG oslo_concurrency.lockutils [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Acquiring lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.084300] env[65107]: DEBUG oslo_concurrency.lockutils [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Acquired lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1250.084473] env[65107]: DEBUG nova.network.neutron [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Refreshing network info cache for port d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1250.360343] env[65107]: DEBUG nova.compute.manager [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1250.360674] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1250.361564] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6edc9b7-ca3f-4739-b029-55226bd1aa24 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.370102] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1250.370370] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b135878-29f0-41d1-997e-d2587e833c35 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.377740] env[65107]: DEBUG oslo_vmware.api [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1250.377740] env[65107]: value = "task-5103867" [ 1250.377740] env[65107]: _type = "Task" [ 1250.377740] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.386634] env[65107]: DEBUG oslo_vmware.api [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.587424] env[65107]: WARNING neutronclient.v2_0.client [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.588223] env[65107]: WARNING openstack [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1250.588621] env[65107]: WARNING openstack [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1250.709804] env[65107]: WARNING openstack [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1250.710275] env[65107]: WARNING openstack [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1250.760879] env[65107]: DEBUG nova.network.neutron [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1250.777105] env[65107]: WARNING neutronclient.v2_0.client [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.777858] env[65107]: WARNING openstack [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1250.778237] env[65107]: WARNING openstack [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1250.867052] env[65107]: DEBUG nova.network.neutron [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updated VIF entry in instance network info cache for port d4ef2ece-8832-4198-a869-6869223cf87a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1250.867457] env[65107]: DEBUG nova.network.neutron [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updating instance_info_cache with network_info: [{"id": "d4ef2ece-8832-4198-a869-6869223cf87a", "address": "fa:16:3e:e0:61:86", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ef2ece-88", "ovs_interfaceid": "d4ef2ece-8832-4198-a869-6869223cf87a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1250.888276] env[65107]: DEBUG oslo_vmware.api [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103867, 'name': PowerOffVM_Task, 'duration_secs': 0.202582} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.888544] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1250.888733] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1250.888996] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e05f9257-8573-4f95-9c0a-9c898a15b7da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.964250] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1250.964491] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1250.964782] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleting the datastore file [datastore2] cd481deb-b12c-47ab-9ab1-0c4bdbb2024a {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1250.965079] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5050fe5-28c5-4432-bcc2-a4248d0965ab {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.972079] env[65107]: DEBUG oslo_vmware.api [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1250.972079] env[65107]: value = "task-5103869" [ 1250.972079] env[65107]: _type = "Task" [ 1250.972079] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.981102] env[65107]: DEBUG oslo_vmware.api [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.263926] env[65107]: INFO nova.compute.manager [-] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Took 1.26 seconds to deallocate network for instance. [ 1251.370759] env[65107]: DEBUG oslo_concurrency.lockutils [req-85522188-b0ac-4f20-974c-1aa65a559701 req-c345bfa7-1480-4c39-9151-81885effd03c service nova] Releasing lock "refresh_cache-e72d0cb3-027c-49af-aa23-b791ae2f5fcf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1251.483013] env[65107]: DEBUG oslo_vmware.api [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144662} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.483326] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1251.483520] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1251.483746] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1251.483942] env[65107]: INFO nova.compute.manager [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1251.484212] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1251.484437] env[65107]: DEBUG nova.compute.manager [-] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1251.484564] env[65107]: DEBUG nova.network.neutron [-] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1251.484838] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1251.485408] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.485774] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.530049] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1251.772524] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1251.772931] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1251.773091] env[65107]: DEBUG nova.objects.instance [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lazy-loading 'resources' on Instance uuid bcced66b-1464-4834-892b-3d0e652da31a {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.118189] env[65107]: DEBUG nova.compute.manager [req-5f67b29c-e183-4b34-9e2b-8dbd85d407d2 req-9860be6b-45e7-4c07-b5c4-fb4881342df9 service nova] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Received event network-vif-deleted-e48b19c2-6253-4a9c-a6fb-1daae3460ede {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1252.118357] env[65107]: DEBUG nova.compute.manager [req-5f67b29c-e183-4b34-9e2b-8dbd85d407d2 req-9860be6b-45e7-4c07-b5c4-fb4881342df9 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Received event network-vif-deleted-cef8867b-7dd6-425a-b5db-760494cf6eb4 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1252.118526] env[65107]: INFO nova.compute.manager [req-5f67b29c-e183-4b34-9e2b-8dbd85d407d2 req-9860be6b-45e7-4c07-b5c4-fb4881342df9 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Neutron deleted interface cef8867b-7dd6-425a-b5db-760494cf6eb4; detaching it from the instance and deleting it from the info cache [ 1252.118734] env[65107]: DEBUG nova.network.neutron [req-5f67b29c-e183-4b34-9e2b-8dbd85d407d2 req-9860be6b-45e7-4c07-b5c4-fb4881342df9 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1252.372149] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0c31c6-7f35-4df7-8393-e38ca3e6cb62 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.381014] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783a74e4-43fb-44a7-83ae-1cc463c8604e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.411893] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295329b2-cc28-4bfe-aa78-dfeefe7952cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.420235] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99db4d16-bbe4-408c-a208-11fbf26f7950 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.435148] env[65107]: DEBUG nova.compute.provider_tree [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.442305] env[65107]: DEBUG nova.network.neutron [-] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1252.621305] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9fab9f3-2b77-478a-94b9-220fa96538dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.631447] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cb7804-efef-4633-8279-623b3d5d1157 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.662091] env[65107]: DEBUG nova.compute.manager [req-5f67b29c-e183-4b34-9e2b-8dbd85d407d2 req-9860be6b-45e7-4c07-b5c4-fb4881342df9 service nova] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Detach interface failed, port_id=cef8867b-7dd6-425a-b5db-760494cf6eb4, reason: Instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1252.939518] env[65107]: DEBUG nova.scheduler.client.report [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1252.944526] env[65107]: INFO nova.compute.manager [-] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Took 1.46 seconds to deallocate network for instance. [ 1253.445820] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.453451] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1253.453780] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.453976] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.478375] env[65107]: INFO nova.scheduler.client.report [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted allocations for instance cd481deb-b12c-47ab-9ab1-0c4bdbb2024a [ 1253.480577] env[65107]: INFO nova.scheduler.client.report [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Deleted allocations for instance bcced66b-1464-4834-892b-3d0e652da31a [ 1253.991754] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a1e449e5-6650-41fb-b765-c3fd2e7dfb25 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "cd481deb-b12c-47ab-9ab1-0c4bdbb2024a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.141s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.992902] env[65107]: DEBUG oslo_concurrency.lockutils [None req-eee918f7-cab3-4dde-a03b-cf938e076c9c tempest-ServersNegativeTestJSON-56640959 tempest-ServersNegativeTestJSON-56640959-project-member] Lock "bcced66b-1464-4834-892b-3d0e652da31a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.639s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1254.919205] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "84f88238-6b21-4530-8a08-72c4049f9150" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1254.919424] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.422924] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1255.860521] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.860782] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.861792] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.861792] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.861792] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.861792] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1255.951506] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.951979] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.953579] env[65107]: INFO nova.compute.claims [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1256.500071] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.039724] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad09c4e-aafb-416a-85ce-590f9e1ef6a4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.048359] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c89c72-fda3-40c5-97bb-65a5b02e8b77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.081082] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a38692d-85db-4e48-9f9e-c84aa861c00f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.089279] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa16f695-afd3-4082-9166-6d6358720701 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.103230] env[65107]: DEBUG nova.compute.provider_tree [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1257.606970] env[65107]: DEBUG nova.scheduler.client.report [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1258.111645] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.160s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1258.112249] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1258.499592] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.617858] env[65107]: DEBUG nova.compute.utils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1258.619448] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1258.619559] env[65107]: DEBUG nova.network.neutron [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1258.619862] env[65107]: WARNING neutronclient.v2_0.client [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.620516] env[65107]: WARNING neutronclient.v2_0.client [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.620997] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.621184] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.672224] env[65107]: DEBUG nova.policy [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6723c29f05774e138ee11fa945c058e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '139a38d3b7bd4d3a8861d2e1e6c7a2c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1258.974167] env[65107]: DEBUG nova.network.neutron [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Successfully created port: ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1259.129917] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1260.140938] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1260.168502] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1260.168764] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1260.168979] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1260.169222] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1260.169371] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1260.169515] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1260.169726] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1260.169913] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1260.170098] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1260.170263] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1260.170430] env[65107]: DEBUG nova.virt.hardware [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1260.171349] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da4e07e-abd5-4e33-aa2a-498e563c99a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.179511] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da0f0d2-9616-47c8-a2fb-806814898955 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.366498] env[65107]: DEBUG nova.compute.manager [req-cd5e3784-417d-4994-9af0-cc76eb036258 req-6e4828c1-d2ca-45f3-b8b2-165aff09ecb2 service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Received event network-vif-plugged-ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1260.366758] env[65107]: DEBUG oslo_concurrency.lockutils [req-cd5e3784-417d-4994-9af0-cc76eb036258 req-6e4828c1-d2ca-45f3-b8b2-165aff09ecb2 service nova] Acquiring lock "84f88238-6b21-4530-8a08-72c4049f9150-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1260.366982] env[65107]: DEBUG oslo_concurrency.lockutils [req-cd5e3784-417d-4994-9af0-cc76eb036258 req-6e4828c1-d2ca-45f3-b8b2-165aff09ecb2 service nova] Lock "84f88238-6b21-4530-8a08-72c4049f9150-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.367215] env[65107]: DEBUG oslo_concurrency.lockutils [req-cd5e3784-417d-4994-9af0-cc76eb036258 req-6e4828c1-d2ca-45f3-b8b2-165aff09ecb2 service nova] Lock "84f88238-6b21-4530-8a08-72c4049f9150-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1260.367428] env[65107]: DEBUG nova.compute.manager [req-cd5e3784-417d-4994-9af0-cc76eb036258 req-6e4828c1-d2ca-45f3-b8b2-165aff09ecb2 service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] No waiting events found dispatching network-vif-plugged-ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1260.367631] env[65107]: WARNING nova.compute.manager [req-cd5e3784-417d-4994-9af0-cc76eb036258 req-6e4828c1-d2ca-45f3-b8b2-165aff09ecb2 service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Received unexpected event network-vif-plugged-ff8925f6-fab7-49a5-bae0-493ec802effc for instance with vm_state building and task_state spawning. [ 1260.463579] env[65107]: DEBUG nova.network.neutron [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Successfully updated port: ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1260.966870] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.966870] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1260.966870] env[65107]: DEBUG nova.network.neutron [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1261.470191] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.470646] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.531448] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.531448] env[65107]: DEBUG nova.network.neutron [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1261.541452] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.541843] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.601764] env[65107]: WARNING neutronclient.v2_0.client [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1261.602451] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1261.602840] env[65107]: WARNING openstack [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1261.681922] env[65107]: DEBUG nova.network.neutron [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1262.003456] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1262.003702] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1262.003872] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1262.004037] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1262.004952] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f69bbca-3083-4bf8-b091-7e305aed8ba5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.013861] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74389994-0237-4782-b824-27806e019d6d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.027743] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d50e2d3-59f9-4722-881f-9cf9e1492d05 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.034428] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7310fc5-a587-41ce-869e-6b842e24d699 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.064983] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179704MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1262.065195] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1262.065370] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1262.185122] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1262.185567] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Instance network_info: |[{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1262.186131] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:0a:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f9ffee1-f413-4f28-8bc4-3fb2cf299789', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff8925f6-fab7-49a5-bae0-493ec802effc', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1262.193422] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1262.193644] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1262.193872] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48f43a44-5f00-4a16-91ac-9cef5d27f75a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.214307] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1262.214307] env[65107]: value = "task-5103875" [ 1262.214307] env[65107]: _type = "Task" [ 1262.214307] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.222203] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103875, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.400047] env[65107]: DEBUG nova.compute.manager [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Received event network-changed-ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1262.400287] env[65107]: DEBUG nova.compute.manager [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Refreshing instance network info cache due to event network-changed-ff8925f6-fab7-49a5-bae0-493ec802effc. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1262.400500] env[65107]: DEBUG oslo_concurrency.lockutils [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.400645] env[65107]: DEBUG oslo_concurrency.lockutils [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1262.400809] env[65107]: DEBUG nova.network.neutron [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Refreshing network info cache for port ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1262.724825] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103875, 'name': CreateVM_Task, 'duration_secs': 0.380803} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.725205] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1262.725515] env[65107]: WARNING neutronclient.v2_0.client [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1262.725889] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.726052] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1262.726366] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1262.726617] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e6cae2d-0f25-4c4e-b9f7-24fc4d7decf5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.731436] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1262.731436] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52240f2b-22c4-389a-42e8-016ab4f568a9" [ 1262.731436] env[65107]: _type = "Task" [ 1262.731436] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.739783] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52240f2b-22c4-389a-42e8-016ab4f568a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.905440] env[65107]: WARNING neutronclient.v2_0.client [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1262.906166] env[65107]: WARNING openstack [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1262.906514] env[65107]: WARNING openstack [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.004723] env[65107]: WARNING openstack [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.005159] env[65107]: WARNING openstack [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.066045] env[65107]: WARNING neutronclient.v2_0.client [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.066698] env[65107]: WARNING openstack [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.067126] env[65107]: WARNING openstack [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.094535] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3df65e64-0c45-4707-960e-8f1767e2d011 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1263.094735] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 751ebb31-1a27-4630-9213-4312a6fa9298 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1263.094865] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 475851f6-24a2-4f0e-9567-d2a23947a5e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1263.094981] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e72d0cb3-027c-49af-aa23-b791ae2f5fcf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1263.095108] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 84f88238-6b21-4530-8a08-72c4049f9150 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1263.095288] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1263.095434] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=100GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '5', 'num_vm_active': '4', 'num_task_None': '4', 'num_os_type_None': '5', 'num_proj_7784762e36ae4c6fb30012ec801c3b88': '2', 'io_workload': '1', 'num_proj_dced4f9aef2f49cf990203b693533aa1': '1', 'num_proj_84acb61dcc764cd79ad2e5b1e2850d16': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1263.156998] env[65107]: DEBUG nova.network.neutron [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updated VIF entry in instance network info cache for port ff8925f6-fab7-49a5-bae0-493ec802effc. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1263.157411] env[65107]: DEBUG nova.network.neutron [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1263.173240] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478b697b-1d42-40be-8a88-ab82230b7f17 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.182604] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975a0546-e544-44b8-9751-ac0325f1608f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.216959] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b8214e-5b4b-447a-8fa7-c715f5045af3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.225108] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca0bd6d-1234-4d9e-8172-9ac25c01366f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.239658] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1263.251182] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52240f2b-22c4-389a-42e8-016ab4f568a9, 'name': SearchDatastore_Task, 'duration_secs': 0.011402} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.251449] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1263.251671] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1263.251900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.252260] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1263.252260] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1263.252785] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e369aef3-7e12-4c56-ad93-46e7457b95e2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.262531] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1263.262682] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1263.263452] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-325e3232-1ba6-442a-94cd-e3d2da3c29f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.269559] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1263.269559] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bc4a08-27ac-33a0-84c2-5133403f1dd0" [ 1263.269559] env[65107]: _type = "Task" [ 1263.269559] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.277696] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bc4a08-27ac-33a0-84c2-5133403f1dd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.661708] env[65107]: DEBUG oslo_concurrency.lockutils [req-92e5e5b5-4b5c-49a0-8315-166fe627b65b req-092609c0-32fc-4a52-a740-dd84b9d1e1cc service nova] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1263.746731] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1263.781982] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52bc4a08-27ac-33a0-84c2-5133403f1dd0, 'name': SearchDatastore_Task, 'duration_secs': 0.009688} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.782566] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-638a447b-51cc-437c-ae98-b84a3e373ad2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.788263] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1263.788263] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52012a35-6c3b-e4cf-dbc9-50f13a498483" [ 1263.788263] env[65107]: _type = "Task" [ 1263.788263] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.796337] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52012a35-6c3b-e4cf-dbc9-50f13a498483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.251621] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1264.251840] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.186s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1264.298586] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52012a35-6c3b-e4cf-dbc9-50f13a498483, 'name': SearchDatastore_Task, 'duration_secs': 0.010617} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.298856] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1264.299211] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1264.299478] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8a5a01f-1835-4f5b-bfe2-238c4e8e6cdb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.306530] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1264.306530] env[65107]: value = "task-5103877" [ 1264.306530] env[65107]: _type = "Task" [ 1264.306530] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.315624] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103877, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.817175] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103877, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493729} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.817635] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1264.817635] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1264.817985] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b5be99c-a59b-4a60-bec1-de0737e2cf0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.825387] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1264.825387] env[65107]: value = "task-5103879" [ 1264.825387] env[65107]: _type = "Task" [ 1264.825387] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.835555] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103879, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.335174] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103879, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0674} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.335446] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1265.336324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9d100a-7e14-435a-8350-0bd9bb6a9ed6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.358669] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfiguring VM instance instance-0000007b to attach disk [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1265.358961] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eba1a831-e916-4885-aaab-8280e9622123 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.379039] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1265.379039] env[65107]: value = "task-5103880" [ 1265.379039] env[65107]: _type = "Task" [ 1265.379039] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.389025] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103880, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.890198] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103880, 'name': ReconfigVM_Task, 'duration_secs': 0.282974} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.890591] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfigured VM instance instance-0000007b to attach disk [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1265.891149] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c22a09ae-644e-4173-bc69-0bdfa47a87ad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.898525] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1265.898525] env[65107]: value = "task-5103881" [ 1265.898525] env[65107]: _type = "Task" [ 1265.898525] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.906975] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103881, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.408499] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103881, 'name': Rename_Task, 'duration_secs': 0.160531} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.408762] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1266.409084] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65fd5de8-03d8-47aa-b0d1-ee50a11fe50c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.416203] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1266.416203] env[65107]: value = "task-5103882" [ 1266.416203] env[65107]: _type = "Task" [ 1266.416203] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.424570] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.926971] env[65107]: DEBUG oslo_vmware.api [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103882, 'name': PowerOnVM_Task, 'duration_secs': 0.500818} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.927361] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1266.927453] env[65107]: INFO nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Took 6.79 seconds to spawn the instance on the hypervisor. [ 1266.927627] env[65107]: DEBUG nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1266.928404] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5644988-a245-47a3-a67f-1804ed083369 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.247025] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.446735] env[65107]: INFO nova.compute.manager [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Took 11.52 seconds to build instance. [ 1267.949241] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ae60e9a8-4dd1-4210-b8ee-75fc77d11d94 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.029s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1268.022030] env[65107]: DEBUG nova.compute.manager [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Received event network-changed-ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1268.022219] env[65107]: DEBUG nova.compute.manager [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Refreshing instance network info cache due to event network-changed-ff8925f6-fab7-49a5-bae0-493ec802effc. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1268.022440] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1268.022715] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1268.022794] env[65107]: DEBUG nova.network.neutron [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Refreshing network info cache for port ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1268.525554] env[65107]: WARNING neutronclient.v2_0.client [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1268.526372] env[65107]: WARNING openstack [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1268.526750] env[65107]: WARNING openstack [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1268.642514] env[65107]: WARNING openstack [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1268.642990] env[65107]: WARNING openstack [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1268.700174] env[65107]: WARNING neutronclient.v2_0.client [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1268.700927] env[65107]: WARNING openstack [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1268.701294] env[65107]: WARNING openstack [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1268.780589] env[65107]: DEBUG nova.network.neutron [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updated VIF entry in instance network info cache for port ff8925f6-fab7-49a5-bae0-493ec802effc. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1268.780961] env[65107]: DEBUG nova.network.neutron [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1269.283924] env[65107]: DEBUG oslo_concurrency.lockutils [req-a8b710aa-e4cb-485c-9f0f-f7df2f931dbe req-9c16ee90-ca4b-436c-9cdd-08a642dae79f service nova] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1270.501675] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1270.502024] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1271.003922] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1271.373466] env[65107]: DEBUG oslo_concurrency.lockutils [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1271.374056] env[65107]: DEBUG oslo_concurrency.lockutils [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1271.527966] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1271.528322] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1271.529974] env[65107]: INFO nova.compute.claims [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1271.877541] env[65107]: INFO nova.compute.manager [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Detaching volume c0acec4c-a8f0-4cbc-a146-0ce2a3362af3 [ 1271.911702] env[65107]: INFO nova.virt.block_device [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Attempting to driver detach volume c0acec4c-a8f0-4cbc-a146-0ce2a3362af3 from mountpoint /dev/sdb [ 1271.911947] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1271.912145] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992903', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'name': 'volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '751ebb31-1a27-4630-9213-4312a6fa9298', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'serial': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1271.913073] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288e3a4d-fc77-4d3e-892d-6ec9845cc993 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.934829] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816593b7-fe35-477b-9465-4faf71f216b5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.945226] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a60bed-fa9c-49b3-91a0-5816715154f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.965742] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b44ef5-39a0-491f-86c4-bee24c65903b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.980901] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The volume has not been displaced from its original location: [datastore2] volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3/volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1271.987221] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Reconfiguring VM instance instance-00000075 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1271.987577] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48cabc09-b550-4a3b-8035-84cfa4ece64c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.007177] env[65107]: DEBUG oslo_vmware.api [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1272.007177] env[65107]: value = "task-5103884" [ 1272.007177] env[65107]: _type = "Task" [ 1272.007177] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.015898] env[65107]: DEBUG oslo_vmware.api [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103884, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.518404] env[65107]: DEBUG oslo_vmware.api [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103884, 'name': ReconfigVM_Task, 'duration_secs': 0.237262} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.518679] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Reconfigured VM instance instance-00000075 to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1272.523416] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-752c3eb2-5ce0-44fd-9b82-e8b3c2108593 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.540555] env[65107]: DEBUG oslo_vmware.api [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1272.540555] env[65107]: value = "task-5103885" [ 1272.540555] env[65107]: _type = "Task" [ 1272.540555] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.549657] env[65107]: DEBUG oslo_vmware.api [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103885, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.624504] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d4ee58-9d19-496a-9dca-513e6d2d47ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.631951] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998f9591-365d-4b8c-8d16-287f69c01dec {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.664234] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6370bc-91ba-480b-9fe5-478b6a6348f9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.672280] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87969ce7-285b-4eb4-bc57-4fcaef5fbbc3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.687285] env[65107]: DEBUG nova.compute.provider_tree [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.052661] env[65107]: DEBUG oslo_vmware.api [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103885, 'name': ReconfigVM_Task, 'duration_secs': 0.159044} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.053055] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992903', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'name': 'volume-c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '751ebb31-1a27-4630-9213-4312a6fa9298', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3', 'serial': 'c0acec4c-a8f0-4cbc-a146-0ce2a3362af3'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1273.190366] env[65107]: DEBUG nova.scheduler.client.report [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1273.593806] env[65107]: DEBUG nova.objects.instance [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1273.695545] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1273.696051] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1274.201061] env[65107]: DEBUG nova.compute.utils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1274.202182] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1274.202480] env[65107]: DEBUG nova.network.neutron [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1274.203326] env[65107]: WARNING neutronclient.v2_0.client [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.203641] env[65107]: WARNING neutronclient.v2_0.client [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.204255] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.204589] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.251229] env[65107]: DEBUG nova.policy [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d55cff40f12420b8320448083ca8f52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7784762e36ae4c6fb30012ec801c3b88', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1274.533156] env[65107]: DEBUG nova.network.neutron [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Successfully created port: 30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1274.542090] env[65107]: DEBUG oslo_concurrency.lockutils [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1274.602403] env[65107]: DEBUG oslo_concurrency.lockutils [None req-194b7643-7280-460c-af8b-840b2dc07644 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.228s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1274.603660] env[65107]: DEBUG oslo_concurrency.lockutils [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.062s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1274.603834] env[65107]: DEBUG nova.compute.manager [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1274.605271] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c22dcf7-051c-4616-b898-fbba5f4e02aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.613039] env[65107]: DEBUG nova.compute.manager [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1274.613756] env[65107]: DEBUG nova.objects.instance [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1274.712853] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1275.218098] env[65107]: INFO nova.virt.block_device [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Booting with volume 56549d08-0e5b-4fd0-adad-3c3b55d4fb93 at /dev/sda [ 1275.253037] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-848a652c-e0e9-47fa-ad1f-6f4825e0ebe0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.263661] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e859dd53-d41a-401c-9e61-336644803336 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.295585] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6d65c65-ed12-4924-9619-35d2c8af494c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.304940] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6603e5e7-ed15-45ed-939e-74561f67b65e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.336165] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6818e7a-98c7-48ce-8559-ecf1744ced4b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.344095] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf4ec9c-f72e-43c1-9fd0-488c2eb91812 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.359650] env[65107]: DEBUG nova.virt.block_device [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating existing volume attachment record: 407df844-afdd-4fd6-9086-fcdbd93cf50f {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1275.620984] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1275.621368] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdb5bc19-12d8-4cf6-bd5c-6d8097f306ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.629720] env[65107]: DEBUG oslo_vmware.api [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1275.629720] env[65107]: value = "task-5103886" [ 1275.629720] env[65107]: _type = "Task" [ 1275.629720] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.639175] env[65107]: DEBUG oslo_vmware.api [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103886, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.930103] env[65107]: DEBUG nova.compute.manager [req-9deb0d91-35cc-41ee-8dfa-9b9cda760997 req-5af5e4ba-9c77-4383-b228-25cfc05d848b service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Received event network-vif-plugged-30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1275.930310] env[65107]: DEBUG oslo_concurrency.lockutils [req-9deb0d91-35cc-41ee-8dfa-9b9cda760997 req-5af5e4ba-9c77-4383-b228-25cfc05d848b service nova] Acquiring lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1275.930548] env[65107]: DEBUG oslo_concurrency.lockutils [req-9deb0d91-35cc-41ee-8dfa-9b9cda760997 req-5af5e4ba-9c77-4383-b228-25cfc05d848b service nova] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1275.930724] env[65107]: DEBUG oslo_concurrency.lockutils [req-9deb0d91-35cc-41ee-8dfa-9b9cda760997 req-5af5e4ba-9c77-4383-b228-25cfc05d848b service nova] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1275.930892] env[65107]: DEBUG nova.compute.manager [req-9deb0d91-35cc-41ee-8dfa-9b9cda760997 req-5af5e4ba-9c77-4383-b228-25cfc05d848b service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] No waiting events found dispatching network-vif-plugged-30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1275.931073] env[65107]: WARNING nova.compute.manager [req-9deb0d91-35cc-41ee-8dfa-9b9cda760997 req-5af5e4ba-9c77-4383-b228-25cfc05d848b service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Received unexpected event network-vif-plugged-30bb95a2-1900-433c-98bc-b8436565a0ff for instance with vm_state building and task_state block_device_mapping. [ 1276.006236] env[65107]: DEBUG nova.network.neutron [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Successfully updated port: 30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1276.140036] env[65107]: DEBUG oslo_vmware.api [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103886, 'name': PowerOffVM_Task, 'duration_secs': 0.203137} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.140245] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1276.140435] env[65107]: DEBUG nova.compute.manager [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1276.141250] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae14585-094a-4648-9be8-5d6169a5b38c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.509825] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1276.509959] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1276.510181] env[65107]: DEBUG nova.network.neutron [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1276.655044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-47866f38-ac61-4b28-9ff1-d5a10cd64b41 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.051s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1277.013431] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.013838] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.047596] env[65107]: DEBUG nova.objects.instance [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.052380] env[65107]: DEBUG nova.network.neutron [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1277.075405] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.075818] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.153357] env[65107]: WARNING neutronclient.v2_0.client [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.154192] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.154551] env[65107]: WARNING openstack [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.252485] env[65107]: DEBUG nova.network.neutron [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1277.448309] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1277.448925] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1277.449161] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1277.449315] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1277.449489] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1277.449628] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1277.449770] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1277.449977] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1277.450181] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1277.450354] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1277.450516] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1277.450685] env[65107]: DEBUG nova.virt.hardware [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1277.451766] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7834ed38-b6c0-49cd-9630-3e513108ab6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.460742] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc17e00c-6801-44d1-916f-15ed40d648aa {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.552855] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1277.553060] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1277.553238] env[65107]: DEBUG nova.network.neutron [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1277.553444] env[65107]: DEBUG nova.objects.instance [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'info_cache' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.755921] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1277.756308] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Instance network_info: |[{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1277.756806] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:be:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30bb95a2-1900-433c-98bc-b8436565a0ff', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1277.764772] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1277.765028] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1277.765270] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b34acce-fe99-4fb7-b2fc-aa3f31d57883 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.785735] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1277.785735] env[65107]: value = "task-5103887" [ 1277.785735] env[65107]: _type = "Task" [ 1277.785735] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.795972] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103887, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.959471] env[65107]: DEBUG nova.compute.manager [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Received event network-changed-30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1277.959616] env[65107]: DEBUG nova.compute.manager [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Refreshing instance network info cache due to event network-changed-30bb95a2-1900-433c-98bc-b8436565a0ff. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1277.959881] env[65107]: DEBUG oslo_concurrency.lockutils [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Acquiring lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1277.960062] env[65107]: DEBUG oslo_concurrency.lockutils [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Acquired lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1277.960257] env[65107]: DEBUG nova.network.neutron [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Refreshing network info cache for port 30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1278.057055] env[65107]: DEBUG nova.objects.base [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Object Instance<751ebb31-1a27-4630-9213-4312a6fa9298> lazy-loaded attributes: flavor,info_cache {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1278.296863] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103887, 'name': CreateVM_Task, 'duration_secs': 0.312565} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.297077] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1278.297514] env[65107]: WARNING neutronclient.v2_0.client [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.297885] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992912', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'name': 'volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6', 'attached_at': '', 'detached_at': '', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'serial': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93'}, 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'attachment_id': '407df844-afdd-4fd6-9086-fcdbd93cf50f', 'mount_device': '/dev/sda', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=65107) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1278.298099] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Root volume attach. Driver type: vmdk {{(pid=65107) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1278.298891] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9798ab6e-5f5e-4489-bbf1-622b780f277c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.306828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd82492-c8e8-44fd-8e97-d41e7ea5b647 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.313017] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18610c1-620f-488d-874d-74fdb4f61f70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.319645] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8f322d6b-eae0-4107-9f51-0d74c78a557a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.327456] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1278.327456] env[65107]: value = "task-5103888" [ 1278.327456] env[65107]: _type = "Task" [ 1278.327456] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.336266] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103888, 'name': RelocateVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.463476] env[65107]: WARNING neutronclient.v2_0.client [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.464206] env[65107]: WARNING openstack [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.464556] env[65107]: WARNING openstack [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.559539] env[65107]: WARNING neutronclient.v2_0.client [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.560294] env[65107]: WARNING openstack [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.560690] env[65107]: WARNING openstack [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.568962] env[65107]: WARNING openstack [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.569225] env[65107]: WARNING openstack [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.632490] env[65107]: WARNING neutronclient.v2_0.client [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.633201] env[65107]: WARNING openstack [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.633562] env[65107]: WARNING openstack [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.692580] env[65107]: WARNING openstack [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.693091] env[65107]: WARNING openstack [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.727150] env[65107]: DEBUG nova.network.neutron [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updated VIF entry in instance network info cache for port 30bb95a2-1900-433c-98bc-b8436565a0ff. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1278.727567] env[65107]: DEBUG nova.network.neutron [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1278.753829] env[65107]: WARNING neutronclient.v2_0.client [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.754031] env[65107]: WARNING openstack [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.754475] env[65107]: WARNING openstack [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.836275] env[65107]: DEBUG nova.network.neutron [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [{"id": "13b14810-c694-4e47-b090-8c3bf39ad510", "address": "fa:16:3e:e7:23:0e", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13b14810-c6", "ovs_interfaceid": "13b14810-c694-4e47-b090-8c3bf39ad510", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1278.840380] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103888, 'name': RelocateVM_Task, 'duration_secs': 0.367095} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.840851] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1278.841066] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992912', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'name': 'volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6', 'attached_at': '', 'detached_at': '', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'serial': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1278.841832] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2bf0e4c-0be7-46e4-8d33-d5be0a9dace5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.858624] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5226c2-39d6-4d27-b7b7-53e8bdbe5720 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.881251] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93/volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1278.881674] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e997865d-5569-42e9-b7d4-04a740489af5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.901917] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1278.901917] env[65107]: value = "task-5103889" [ 1278.901917] env[65107]: _type = "Task" [ 1278.901917] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.913204] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.230485] env[65107]: DEBUG oslo_concurrency.lockutils [req-9c147b8a-f7d5-4c9c-a630-557c48f441a2 req-c2dec493-6c0b-423d-b443-b92fd83c05c3 service nova] Releasing lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1279.342190] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "refresh_cache-751ebb31-1a27-4630-9213-4312a6fa9298" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1279.412833] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103889, 'name': ReconfigVM_Task, 'duration_secs': 0.248123} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.413150] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfigured VM instance instance-0000007c to attach disk [datastore2] volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93/volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1279.418426] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80c76fdf-61d3-429e-98fa-5b4c485fdd91 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.435612] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1279.435612] env[65107]: value = "task-5103890" [ 1279.435612] env[65107]: _type = "Task" [ 1279.435612] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.444923] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103890, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.946996] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103890, 'name': ReconfigVM_Task, 'duration_secs': 0.128594} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.947477] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992912', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'name': 'volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6', 'attached_at': '', 'detached_at': '', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'serial': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1279.947988] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-410c7b7a-597f-41c7-ac08-5ce7d08909ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.956019] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1279.956019] env[65107]: value = "task-5103891" [ 1279.956019] env[65107]: _type = "Task" [ 1279.956019] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.965268] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103891, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.348054] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1280.348054] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18075cc6-724c-4ea2-b125-b1e2fbe5d04f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.356235] env[65107]: DEBUG oslo_vmware.api [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1280.356235] env[65107]: value = "task-5103892" [ 1280.356235] env[65107]: _type = "Task" [ 1280.356235] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.364840] env[65107]: DEBUG oslo_vmware.api [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.465943] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103891, 'name': Rename_Task, 'duration_secs': 0.124879} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.466325] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1280.466564] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e6530cb-3ae0-43ec-a261-3aa512f64857 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.473094] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1280.473094] env[65107]: value = "task-5103893" [ 1280.473094] env[65107]: _type = "Task" [ 1280.473094] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.481632] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.867909] env[65107]: DEBUG oslo_vmware.api [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103892, 'name': PowerOnVM_Task, 'duration_secs': 0.378959} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.868220] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1280.868457] env[65107]: DEBUG nova.compute.manager [None req-7d3558d2-a931-4c77-8772-c99f283d5ee9 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1280.869308] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd7c9c8-014c-46b7-b3f3-99253d94a981 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.983793] env[65107]: DEBUG oslo_vmware.api [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103893, 'name': PowerOnVM_Task, 'duration_secs': 0.452305} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.984277] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1280.984347] env[65107]: INFO nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Took 3.54 seconds to spawn the instance on the hypervisor. [ 1280.984508] env[65107]: DEBUG nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1280.985375] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e3602c-1965-4bee-a7b3-5d52b85b8917 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.503881] env[65107]: INFO nova.compute.manager [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Took 9.99 seconds to build instance. [ 1282.005449] env[65107]: DEBUG oslo_concurrency.lockutils [None req-75ad3ae7-6033-474a-96c3-993acea5790a tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.503s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1283.249642] env[65107]: DEBUG nova.compute.manager [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Received event network-changed-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1283.249642] env[65107]: DEBUG nova.compute.manager [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Refreshing instance network info cache due to event network-changed-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1283.250021] env[65107]: DEBUG oslo_concurrency.lockutils [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Acquiring lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.250021] env[65107]: DEBUG oslo_concurrency.lockutils [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Acquired lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.250021] env[65107]: DEBUG nova.network.neutron [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Refreshing network info cache for port 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1283.753606] env[65107]: WARNING neutronclient.v2_0.client [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.754367] env[65107]: WARNING openstack [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.755041] env[65107]: WARNING openstack [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.874815] env[65107]: WARNING openstack [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.875340] env[65107]: WARNING openstack [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.938830] env[65107]: WARNING neutronclient.v2_0.client [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.939519] env[65107]: WARNING openstack [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.939862] env[65107]: WARNING openstack [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.026080] env[65107]: DEBUG nova.network.neutron [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updated VIF entry in instance network info cache for port 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1284.026519] env[65107]: DEBUG nova.network.neutron [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updating instance_info_cache with network_info: [{"id": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "address": "fa:16:3e:4e:cd:29", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75cfc6b6-0a", "ovs_interfaceid": "75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1284.530024] env[65107]: DEBUG oslo_concurrency.lockutils [req-26c8912e-6641-4d97-9aa0-bdf8f23a1c4a req-d25a59b2-a75d-4132-9e14-1a39630bafda service nova] Releasing lock "refresh_cache-3df65e64-0c45-4707-960e-8f1767e2d011" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.734903] env[65107]: DEBUG nova.compute.manager [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1284.863215] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1284.863483] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1284.863662] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1284.863909] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1284.864130] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1284.866617] env[65107]: INFO nova.compute.manager [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Terminating instance [ 1285.252963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1285.253792] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1285.281677] env[65107]: DEBUG nova.compute.manager [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Received event network-changed-30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1285.281876] env[65107]: DEBUG nova.compute.manager [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Refreshing instance network info cache due to event network-changed-30bb95a2-1900-433c-98bc-b8436565a0ff. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1285.282139] env[65107]: DEBUG oslo_concurrency.lockutils [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Acquiring lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.282293] env[65107]: DEBUG oslo_concurrency.lockutils [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Acquired lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1285.282453] env[65107]: DEBUG nova.network.neutron [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Refreshing network info cache for port 30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1285.370190] env[65107]: DEBUG nova.compute.manager [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1285.370430] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1285.371441] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f24c28-f615-4615-b27d-a096027201e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.380523] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1285.380780] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26afffad-203e-4a19-a1ed-128c201d9e55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.387999] env[65107]: DEBUG oslo_vmware.api [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1285.387999] env[65107]: value = "task-5103894" [ 1285.387999] env[65107]: _type = "Task" [ 1285.387999] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.397722] env[65107]: DEBUG oslo_vmware.api [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103894, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.758866] env[65107]: INFO nova.compute.claims [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1285.784988] env[65107]: WARNING neutronclient.v2_0.client [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1285.785690] env[65107]: WARNING openstack [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.786076] env[65107]: WARNING openstack [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.639911] env[65107]: INFO nova.compute.resource_tracker [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating resource usage from migration 3673c926-919b-4e1e-a0aa-144d7536d6b2 [ 1286.653436] env[65107]: DEBUG oslo_vmware.api [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103894, 'name': PowerOffVM_Task, 'duration_secs': 0.241878} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.653748] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1286.653913] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1286.654190] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf6f2d12-f05f-4ac3-9d3d-99cba9684cb3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.726127] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1286.726439] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1286.726663] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleting the datastore file [datastore2] 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1286.727016] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e56e988-d1ea-42e6-8e8b-f5b8ef185cd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.730822] env[65107]: WARNING openstack [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.731270] env[65107]: WARNING openstack [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.747955] env[65107]: DEBUG oslo_vmware.api [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for the task: (returnval){ [ 1286.747955] env[65107]: value = "task-5103896" [ 1286.747955] env[65107]: _type = "Task" [ 1286.747955] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.764647] env[65107]: DEBUG oslo_vmware.api [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.768731] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84bc74c-b557-4d42-b78d-b02ef2cb1117 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.780061] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbf0dfb-eadd-4f3b-9527-bd1bebdd2082 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.815167] env[65107]: WARNING neutronclient.v2_0.client [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.815923] env[65107]: WARNING openstack [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.816329] env[65107]: WARNING openstack [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.825202] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47042aba-2d6a-4ae7-898e-5821c8bb24bf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.834683] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca399c6f-f3db-4bde-a94f-6e6d3b462bb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.848551] env[65107]: DEBUG nova.compute.provider_tree [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1286.908976] env[65107]: DEBUG nova.network.neutron [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updated VIF entry in instance network info cache for port 30bb95a2-1900-433c-98bc-b8436565a0ff. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1286.909355] env[65107]: DEBUG nova.network.neutron [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1287.260061] env[65107]: DEBUG oslo_vmware.api [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Task: {'id': task-5103896, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14386} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.260061] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1287.260061] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1287.260321] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1287.260385] env[65107]: INFO nova.compute.manager [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Took 1.89 seconds to destroy the instance on the hypervisor. [ 1287.260630] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1287.260820] env[65107]: DEBUG nova.compute.manager [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1287.260916] env[65107]: DEBUG nova.network.neutron [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1287.261172] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1287.261675] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1287.261923] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1287.298304] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1287.352221] env[65107]: DEBUG nova.scheduler.client.report [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1287.412097] env[65107]: DEBUG oslo_concurrency.lockutils [req-a32c043f-90aa-46de-840e-b89741c3ed99 req-bd66df1c-4ba7-42c5-bc1f-55a6bf09d5fd service nova] Releasing lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1287.713121] env[65107]: DEBUG nova.compute.manager [req-20b50b27-858e-4e6e-b6a3-cbfccca525cb req-1c4cc03c-9a6b-448a-bcef-7bb57119109c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Received event network-vif-deleted-72f482ac-0a93-4cf4-9d54-a7e812ca4cd9 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1287.713481] env[65107]: INFO nova.compute.manager [req-20b50b27-858e-4e6e-b6a3-cbfccca525cb req-1c4cc03c-9a6b-448a-bcef-7bb57119109c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Neutron deleted interface 72f482ac-0a93-4cf4-9d54-a7e812ca4cd9; detaching it from the instance and deleting it from the info cache [ 1287.714250] env[65107]: DEBUG nova.network.neutron [req-20b50b27-858e-4e6e-b6a3-cbfccca525cb req-1c4cc03c-9a6b-448a-bcef-7bb57119109c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1287.856988] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.604s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1287.857379] env[65107]: INFO nova.compute.manager [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Migrating [ 1288.191115] env[65107]: DEBUG nova.network.neutron [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1288.215703] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8dc61b3d-9d9c-4730-a9bc-f2b51e1fa84c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.228677] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce0c71e-dff4-42ef-9042-5b0b940d4dc0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.267373] env[65107]: DEBUG nova.compute.manager [req-20b50b27-858e-4e6e-b6a3-cbfccca525cb req-1c4cc03c-9a6b-448a-bcef-7bb57119109c service nova] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Detach interface failed, port_id=72f482ac-0a93-4cf4-9d54-a7e812ca4cd9, reason: Instance 475851f6-24a2-4f0e-9567-d2a23947a5e3 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1288.372524] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.372766] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1288.372960] env[65107]: DEBUG nova.network.neutron [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1288.694092] env[65107]: INFO nova.compute.manager [-] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Took 1.43 seconds to deallocate network for instance. [ 1288.876442] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1288.877168] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.877568] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1289.003654] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1289.004099] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1289.076076] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1289.076742] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1289.077104] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1289.155485] env[65107]: DEBUG nova.network.neutron [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1289.200985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.201281] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.201499] env[65107]: DEBUG nova.objects.instance [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lazy-loading 'resources' on Instance uuid 475851f6-24a2-4f0e-9567-d2a23947a5e3 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1289.658365] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1289.801362] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082ad8da-4ac9-426f-b49e-dc887c543f96 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.809979] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799ad3cb-6f07-4480-9ffb-efe306933f78 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.840052] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db87d3af-ab2d-4310-bf7b-4085b0d136f5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.847873] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecba20b3-9c30-46da-af94-f925567d8b9b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.861367] env[65107]: DEBUG nova.compute.provider_tree [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1290.364345] env[65107]: DEBUG nova.scheduler.client.report [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1290.870982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.670s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1290.888406] env[65107]: INFO nova.scheduler.client.report [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Deleted allocations for instance 475851f6-24a2-4f0e-9567-d2a23947a5e3 [ 1291.174400] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8bef069-92c4-489d-9b0d-a4b90d93c198 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.193990] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1291.396092] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fe1514ca-8240-4805-bf97-03dba651c254 tempest-AttachVolumeShelveTestJSON-733143698 tempest-AttachVolumeShelveTestJSON-733143698-project-member] Lock "475851f6-24a2-4f0e-9567-d2a23947a5e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.532s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1291.701082] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1291.701480] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9928af6e-3351-48d6-b107-4f0ff445a5e6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.710420] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1291.710420] env[65107]: value = "task-5103897" [ 1291.710420] env[65107]: _type = "Task" [ 1291.710420] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.719765] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.220743] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103897, 'name': PowerOffVM_Task, 'duration_secs': 0.187711} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.220916] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1292.221069] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1292.727519] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1292.727981] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1292.727981] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1292.728090] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1292.728231] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1292.728378] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1292.728643] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1292.728791] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1292.728920] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1292.729099] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1292.729387] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1292.734410] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-702cde29-ccab-48a0-8abe-a2e2d0d97690 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.750157] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1292.750157] env[65107]: value = "task-5103899" [ 1292.750157] env[65107]: _type = "Task" [ 1292.750157] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.758899] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103899, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.260720] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103899, 'name': ReconfigVM_Task, 'duration_secs': 0.169627} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.261020] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1293.767116] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1293.767570] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1293.767622] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1293.767830] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1293.767980] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1293.768150] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1293.768365] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1293.768596] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1293.768785] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1293.768952] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1293.769168] env[65107]: DEBUG nova.virt.hardware [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1293.774562] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfiguring VM instance instance-0000007c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1293.774871] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ac424d9-733c-40f8-af77-c43297e76c7c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.797770] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1293.797770] env[65107]: value = "task-5103900" [ 1293.797770] env[65107]: _type = "Task" [ 1293.797770] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.806717] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103900, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.307901] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103900, 'name': ReconfigVM_Task, 'duration_secs': 0.163985} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.308200] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfigured VM instance instance-0000007c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1294.308990] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e837538a-9f05-4d84-ab6c-a9a6a02deb3c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.330591] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93/volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1294.330920] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5bb75a8-8a38-4df8-9034-954f665dfb09 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.348787] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1294.348787] env[65107]: value = "task-5103901" [ 1294.348787] env[65107]: _type = "Task" [ 1294.348787] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.357043] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103901, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.858720] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103901, 'name': ReconfigVM_Task, 'duration_secs': 0.242326} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.860043] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfigured VM instance instance-0000007c to attach disk [datastore2] volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93/volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1294.860043] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1295.366336] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39481eb2-175d-45ed-b4c3-e874bd655983 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.387077] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f802d55-8460-4639-99fc-3b05ed8e0047 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.406445] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1297.021141] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1297.021617] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1297.077890] env[65107]: DEBUG nova.network.neutron [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Port 30bb95a2-1900-433c-98bc-b8436565a0ff binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1298.100236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1298.100236] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1298.100612] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1299.105641] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.144613] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.144791] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1299.144967] env[65107]: DEBUG nova.network.neutron [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1299.647956] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.648717] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1299.649082] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.780057] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1299.780745] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.849836] env[65107]: WARNING neutronclient.v2_0.client [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.850606] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1299.850938] env[65107]: WARNING openstack [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.942477] env[65107]: DEBUG nova.network.neutron [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1300.445190] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1300.954603] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf6d78c-dc44-4b06-ad78-f390864d59c9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.962199] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261688b8-fc55-4f01-a470-7afe420a0b59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.063317] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70c2b68-f0f8-4ebc-b7c1-dd0691bea32a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.084174] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad96422-e2a1-484c-b9a8-72ed44500ab4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.091759] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1302.598434] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1302.599364] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08f6be36-7c08-47e7-8202-e89120dc9ec1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.606760] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1302.606760] env[65107]: value = "task-5103902" [ 1302.606760] env[65107]: _type = "Task" [ 1302.606760] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.615423] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103902, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.118077] env[65107]: DEBUG oslo_vmware.api [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103902, 'name': PowerOnVM_Task, 'duration_secs': 0.385916} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.118077] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1303.118077] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-7b59ca5e-58b6-4b44-9007-26853dfc3135 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1305.422026] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1305.422440] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1305.422520] env[65107]: DEBUG nova.compute.manager [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Going to confirm migration 9 {{(pid=65107) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1305.929072] env[65107]: WARNING neutronclient.v2_0.client [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1305.962630] env[65107]: WARNING neutronclient.v2_0.client [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1305.963008] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.963178] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquired lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1305.963353] env[65107]: DEBUG nova.network.neutron [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1305.963535] env[65107]: DEBUG nova.objects.instance [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'info_cache' on Instance uuid f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1306.194579] env[65107]: DEBUG nova.compute.manager [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Stashing vm_state: active {{(pid=65107) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1306.717017] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.717333] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.969799] env[65107]: WARNING neutronclient.v2_0.client [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.970558] env[65107]: WARNING openstack [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.970905] env[65107]: WARNING openstack [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.090128] env[65107]: WARNING openstack [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.090540] env[65107]: WARNING openstack [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.151571] env[65107]: WARNING neutronclient.v2_0.client [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1307.152658] env[65107]: WARNING openstack [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.152658] env[65107]: WARNING openstack [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.222634] env[65107]: INFO nova.compute.claims [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1307.232137] env[65107]: DEBUG nova.network.neutron [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [{"id": "30bb95a2-1900-433c-98bc-b8436565a0ff", "address": "fa:16:3e:f7:be:c2", "network": {"id": "731b472f-dbc3-4cb9-aa3c-f9f3f116c19f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1386584066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7784762e36ae4c6fb30012ec801c3b88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb95a2-19", "ovs_interfaceid": "30bb95a2-1900-433c-98bc-b8436565a0ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1307.729721] env[65107]: INFO nova.compute.resource_tracker [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating resource usage from migration 2c5a4459-df77-4ef8-8722-39e760c95450 [ 1307.735625] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Releasing lock "refresh_cache-f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1307.735873] env[65107]: DEBUG nova.objects.instance [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'migration_context' on Instance uuid f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1307.817935] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db321f9-2c25-4718-be18-e2c04573959b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.825613] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bca6cd-4193-4549-9fb2-7d419d9d46b2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.855012] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228e15c8-448b-436b-bc35-44d41858444a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.862542] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1579b70-7cb3-49dc-af95-00807026ada0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.876983] env[65107]: DEBUG nova.compute.provider_tree [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1308.239179] env[65107]: DEBUG nova.objects.base [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1308.240135] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b5c063-13ee-4e36-af5f-1308e7d65fad {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.259522] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-001cc72d-b3d0-4bb8-b64c-f4e8fa17c6bc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.265238] env[65107]: DEBUG oslo_vmware.api [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1308.265238] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eb0d76-a6c5-7754-1de2-323f2b82d61b" [ 1308.265238] env[65107]: _type = "Task" [ 1308.265238] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.273115] env[65107]: DEBUG oslo_vmware.api [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eb0d76-a6c5-7754-1de2-323f2b82d61b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.379984] env[65107]: DEBUG nova.scheduler.client.report [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1308.776490] env[65107]: DEBUG oslo_vmware.api [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52eb0d76-a6c5-7754-1de2-323f2b82d61b, 'name': SearchDatastore_Task, 'duration_secs': 0.009805} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.776865] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1308.885072] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.168s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1308.885327] env[65107]: INFO nova.compute.manager [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Migrating [ 1308.891906] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.115s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1309.400733] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.400989] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1309.401094] env[65107]: DEBUG nova.network.neutron [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1309.501947] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e06665-7916-420e-8df4-a97429ab8b25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.510116] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7147e835-3c1e-47ad-b6b4-bbd93a7d2fc7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.541688] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98e5c88-417f-40eb-bcf6-60cad8bd461e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.549945] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe8a6ca-fdf5-429f-83e4-78fa9b774a34 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.564205] env[65107]: DEBUG nova.compute.provider_tree [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1309.906771] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1309.907451] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1309.907856] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1310.024672] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1310.025080] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1310.067575] env[65107]: DEBUG nova.scheduler.client.report [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1310.086771] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1310.087439] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1310.087790] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1310.169675] env[65107]: DEBUG nova.network.neutron [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1310.672869] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1311.078545] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.186s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1311.639868] env[65107]: INFO nova.scheduler.client.report [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted allocation for migration 3673c926-919b-4e1e-a0aa-144d7536d6b2 [ 1311.906564] env[65107]: INFO nova.compute.manager [None req-56dde37f-e771-4b9b-bf48-959ea89a5823 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Get console output [ 1311.907081] env[65107]: WARNING nova.virt.vmwareapi.driver [None req-56dde37f-e771-4b9b-bf48-959ea89a5823 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] The console log is missing. Check your VSPC configuration [ 1312.145747] env[65107]: DEBUG oslo_concurrency.lockutils [None req-30777704-90b0-4d15-8df8-d4a02372271b tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.723s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.188253] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1d125a-9978-4b3e-ada3-9d856a60160e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.207346] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 0 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1312.713391] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1312.713727] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44485891-c783-452c-8c14-025bc5931bc4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.721713] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1312.721713] env[65107]: value = "task-5103903" [ 1312.721713] env[65107]: _type = "Task" [ 1312.721713] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.730522] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103903, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.232287] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103903, 'name': PowerOffVM_Task, 'duration_secs': 0.186384} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.232597] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1313.232751] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 17 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1313.499544] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1313.741033] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1313.741033] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1313.741033] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1313.741033] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1313.741033] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1313.741568] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1313.741974] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1313.742728] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1313.742728] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1313.742728] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1313.742904] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1313.748360] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7fef57a-f171-466a-8949-50db691e43d0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.764751] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1313.764751] env[65107]: value = "task-5103904" [ 1313.764751] env[65107]: _type = "Task" [ 1313.764751] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.775373] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103904, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.275413] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103904, 'name': ReconfigVM_Task, 'duration_secs': 0.169311} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.275900] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 33 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1314.495216] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1314.782513] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1314.782797] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1314.782900] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1314.783094] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1314.783242] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1314.783415] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1314.783644] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1314.783819] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1314.783989] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1314.784168] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1314.784340] env[65107]: DEBUG nova.virt.hardware [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1314.789645] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfiguring VM instance instance-0000007b to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1314.789943] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-270cba11-6119-4f7c-9b0d-8bd9e8fa5030 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.808788] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1314.808788] env[65107]: value = "task-5103905" [ 1314.808788] env[65107]: _type = "Task" [ 1314.808788] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.819247] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103905, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.319239] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103905, 'name': ReconfigVM_Task, 'duration_secs': 0.165237} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.319645] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfigured VM instance instance-0000007b to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1315.320292] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3d4057-d0e9-42ce-95e2-ad3d3d65464a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.342819] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfiguring VM instance instance-0000007b to attach disk [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1315.343117] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4d902d5-b42c-480f-8d87-d3af6d31ea73 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.361340] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1315.361340] env[65107]: value = "task-5103906" [ 1315.361340] env[65107]: _type = "Task" [ 1315.361340] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.369676] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103906, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.500557] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.500857] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.501058] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1315.872478] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.373672] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103906, 'name': ReconfigVM_Task, 'duration_secs': 0.796895} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.374160] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfigured VM instance instance-0000007b to attach disk [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1316.374263] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 50 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1316.500321] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1316.880995] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56f8e21-6b37-4d23-9f99-bbae7fb0f23b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.901850] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c28ee4-8dd6-45aa-8b53-bf73ea5c988e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.923652] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 67 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1317.430044] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1317.430503] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1317.481671] env[65107]: DEBUG nova.network.neutron [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Port ff8925f6-fab7-49a5-bae0-493ec802effc binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1317.499681] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.521218] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1317.521596] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1317.521819] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "751ebb31-1a27-4630-9213-4312a6fa9298-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1317.522012] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1317.522188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1317.524496] env[65107]: INFO nova.compute.manager [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Terminating instance [ 1318.028936] env[65107]: DEBUG nova.compute.manager [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1318.029165] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1318.030121] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e89115-b933-4a50-9136-7849c9fdceb7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.038548] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1318.038808] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd98c74e-8d73-4bcf-953e-1f1bb7530c64 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.046591] env[65107]: DEBUG oslo_vmware.api [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1318.046591] env[65107]: value = "task-5103907" [ 1318.046591] env[65107]: _type = "Task" [ 1318.046591] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.057074] env[65107]: DEBUG oslo_vmware.api [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.500526] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.502425] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "84f88238-6b21-4530-8a08-72c4049f9150-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1318.502650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1318.502821] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1318.557782] env[65107]: DEBUG oslo_vmware.api [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103907, 'name': PowerOffVM_Task, 'duration_secs': 0.21066} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1318.558033] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1318.558185] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1318.558441] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ef05a0d-c5d9-477b-87dd-6bae3327dfda {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.631270] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1318.631558] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1318.631775] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Deleting the datastore file [datastore1] 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1318.632050] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-91140309-33df-4c5f-8109-3fe71be99dcc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.639384] env[65107]: DEBUG oslo_vmware.api [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1318.639384] env[65107]: value = "task-5103909" [ 1318.639384] env[65107]: _type = "Task" [ 1318.639384] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.649840] env[65107]: DEBUG oslo_vmware.api [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.150517] env[65107]: DEBUG oslo_vmware.api [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151225} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.150879] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1319.151182] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1319.151474] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1319.151754] env[65107]: INFO nova.compute.manager [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1319.152145] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1319.152431] env[65107]: DEBUG nova.compute.manager [-] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1319.152590] env[65107]: DEBUG nova.network.neutron [-] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1319.152945] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1319.153831] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1319.154295] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1319.195499] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1319.512285] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1319.542242] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.542422] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1319.542595] env[65107]: DEBUG nova.network.neutron [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1319.627295] env[65107]: DEBUG nova.compute.manager [req-9eb87c0d-cce8-4de3-b733-89aa13208808 req-ae560cdc-51ba-40a5-8e34-7eca8f2845f5 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Received event network-vif-deleted-13b14810-c694-4e47-b090-8c3bf39ad510 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1319.627295] env[65107]: INFO nova.compute.manager [req-9eb87c0d-cce8-4de3-b733-89aa13208808 req-ae560cdc-51ba-40a5-8e34-7eca8f2845f5 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Neutron deleted interface 13b14810-c694-4e47-b090-8c3bf39ad510; detaching it from the instance and deleting it from the info cache [ 1319.627447] env[65107]: DEBUG nova.network.neutron [req-9eb87c0d-cce8-4de3-b733-89aa13208808 req-ae560cdc-51ba-40a5-8e34-7eca8f2845f5 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1320.045467] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1320.046525] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.046769] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.113080] env[65107]: DEBUG nova.network.neutron [-] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1320.130645] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d015f8f6-8808-494f-9848-56f3a0db7277 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.141376] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8cd9e3-80f4-432f-b5a0-71a86a0cfbd1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.162488] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.162884] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.179172] env[65107]: DEBUG nova.compute.manager [req-9eb87c0d-cce8-4de3-b733-89aa13208808 req-ae560cdc-51ba-40a5-8e34-7eca8f2845f5 service nova] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Detach interface failed, port_id=13b14810-c694-4e47-b090-8c3bf39ad510, reason: Instance 751ebb31-1a27-4630-9213-4312a6fa9298 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1320.235654] env[65107]: WARNING neutronclient.v2_0.client [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1320.236380] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.236919] env[65107]: WARNING openstack [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.330834] env[65107]: DEBUG nova.network.neutron [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1320.615611] env[65107]: INFO nova.compute.manager [-] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Took 1.46 seconds to deallocate network for instance. [ 1320.833953] env[65107]: DEBUG oslo_concurrency.lockutils [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1321.122187] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1321.122503] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.122749] env[65107]: DEBUG nova.objects.instance [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'resources' on Instance uuid 751ebb31-1a27-4630-9213-4312a6fa9298 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1321.363207] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515b7d25-554d-4ae1-92b2-e94b4a2da001 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.384357] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8689c60d-af33-4b30-9c56-85cf6fcee6f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.392602] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 83 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1321.708052] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430be215-9bfd-42e8-82c9-1de314d116ff {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.714977] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f898d10-fc7c-4e1b-add6-d0acfa799237 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.744428] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e76c75-9df4-431f-95b7-a91f5406181c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.752081] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa89dd9-e507-4dd1-95c2-2e8eff44f367 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.765735] env[65107]: DEBUG nova.compute.provider_tree [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1321.899362] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1321.899681] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e6df854-f1dc-401e-a23a-9f858937a214 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.907285] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1321.907285] env[65107]: value = "task-5103910" [ 1321.907285] env[65107]: _type = "Task" [ 1321.907285] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.915522] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103910, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.269606] env[65107]: DEBUG nova.scheduler.client.report [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1322.419389] env[65107]: DEBUG oslo_vmware.api [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103910, 'name': PowerOnVM_Task, 'duration_secs': 0.39118} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.419604] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1322.419792] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-89017a51-01c6-475c-b895-73cab80cd8f8 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance '84f88238-6b21-4530-8a08-72c4049f9150' progress to 100 {{(pid=65107) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1322.776928] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1322.799784] env[65107]: INFO nova.scheduler.client.report [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Deleted allocations for instance 751ebb31-1a27-4630-9213-4312a6fa9298 [ 1323.308853] env[65107]: DEBUG oslo_concurrency.lockutils [None req-94279f25-cca8-4acc-bb3f-b88df856a1dc tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "751ebb31-1a27-4630-9213-4312a6fa9298" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.787s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1323.499532] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.003206] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1324.003548] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1324.003639] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1324.003798] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1324.004733] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642b0c7a-9be9-4e22-bee1-02371a4bf98d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.013732] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a068d5f4-6ae7-47a1-9c68-6905f59844bb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.029472] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e380df9d-3cac-469e-b3b7-824f8d9bd475 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.036935] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcbb809-32ce-4f07-914a-cbb8cfedd9d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.067688] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179549MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1324.067845] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1324.068078] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1325.079473] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Applying migration context for instance 84f88238-6b21-4530-8a08-72c4049f9150 as it has an incoming, in-progress migration 2c5a4459-df77-4ef8-8722-39e760c95450. Migration status is reverting {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1325.080307] env[65107]: INFO nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating resource usage from migration 2c5a4459-df77-4ef8-8722-39e760c95450 [ 1325.092981] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.099171] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 3df65e64-0c45-4707-960e-8f1767e2d011 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1325.099349] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance e72d0cb3-027c-49af-aa23-b791ae2f5fcf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1325.099441] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1325.099557] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Migration 2c5a4459-df77-4ef8-8722-39e760c95450 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1325.099670] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 84f88238-6b21-4530-8a08-72c4049f9150 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1325.099852] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1325.099992] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1600MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=5 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '3', 'num_task_None': '4', 'num_os_type_None': '4', 'num_proj_7784762e36ae4c6fb30012ec801c3b88': '3', 'io_workload': '0', 'num_vm_resized': '1', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1325.135430] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.135981] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.136308] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.170618] env[65107]: DEBUG nova.network.neutron [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Port ff8925f6-fab7-49a5-bae0-493ec802effc binding to destination host cpu-1 is already ACTIVE {{(pid=65107) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3279}} [ 1325.170893] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1325.171058] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1325.171227] env[65107]: DEBUG nova.network.neutron [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1325.174013] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b8b7e7-7cba-4332-bebf-5e1d0cb79bd7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.183438] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3d5525-b393-450b-bef2-653c4f435956 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.213828] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48c567b-19b7-41e6-a3b3-96b2d42e800b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.221810] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c9dbbf-6637-4229-bf71-d8f0a831090b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.237939] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1325.678498] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.679587] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1325.679951] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1325.693476] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1325.693824] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1325.740854] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1325.812251] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1325.812668] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1325.874250] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.874889] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1325.875249] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1325.955642] env[65107]: DEBUG nova.network.neutron [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1326.196055] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1326.246750] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1326.247233] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.179s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1326.459370] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1326.719645] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1326.719936] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1326.721981] env[65107]: INFO nova.compute.claims [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1326.963888] env[65107]: DEBUG nova.compute.manager [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65107) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1326.964132] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1327.811468] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee778fc-dc93-4d2c-9c01-86e90b172f22 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.819707] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d29115-b3d0-4132-8ee1-7f0221c51c39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.851219] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f55949b-af42-4d29-970f-d64c5eafeb3a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.859094] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71cb311-c7bb-4025-a50b-bf32c9c939e8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.872342] env[65107]: DEBUG nova.compute.provider_tree [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1328.376279] env[65107]: DEBUG nova.scheduler.client.report [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1328.881145] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.161s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1328.881657] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1328.884388] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 1.920s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1329.388276] env[65107]: DEBUG nova.compute.utils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1329.389821] env[65107]: DEBUG nova.objects.instance [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'migration_context' on Instance uuid 84f88238-6b21-4530-8a08-72c4049f9150 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1329.391012] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1329.391214] env[65107]: DEBUG nova.network.neutron [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1329.391544] env[65107]: WARNING neutronclient.v2_0.client [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1329.391856] env[65107]: WARNING neutronclient.v2_0.client [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1329.392445] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1329.392784] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1329.437407] env[65107]: DEBUG nova.policy [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5fb11f9a7f72499eb60964784b8b9afe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84acb61dcc764cd79ad2e5b1e2850d16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1329.486895] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adf570f-f9a7-4495-8bfb-e82b36c5329f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.495171] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38f3344-7a1d-4575-b62d-eee2405cbe5d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.527456] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5c52df-86b0-41b0-9fda-d49af286fdb8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.536136] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9842e054-9a2d-4a53-b086-5afa97dc4192 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.551308] env[65107]: DEBUG nova.compute.provider_tree [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1329.736970] env[65107]: DEBUG nova.network.neutron [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Successfully created port: a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1329.892294] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1330.054291] env[65107]: DEBUG nova.scheduler.client.report [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1330.902622] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1330.929509] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1330.929816] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1330.930021] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1330.930268] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1330.930453] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1330.930642] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1330.930930] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1330.931178] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1330.931777] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1330.931777] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1330.931907] env[65107]: DEBUG nova.virt.hardware [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1330.932791] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4d7494-ba77-439b-8255-6cc20b741411 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.941867] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e16419-c878-4293-9aab-f7227ece3d06 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.065691] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.181s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1331.173547] env[65107]: DEBUG nova.compute.manager [req-1218d9e1-51d3-4dcc-9030-46c8afb78842 req-e986b1fa-838b-4d58-8039-e9227e7ba1c6 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Received event network-vif-plugged-a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1331.173706] env[65107]: DEBUG oslo_concurrency.lockutils [req-1218d9e1-51d3-4dcc-9030-46c8afb78842 req-e986b1fa-838b-4d58-8039-e9227e7ba1c6 service nova] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1331.174061] env[65107]: DEBUG oslo_concurrency.lockutils [req-1218d9e1-51d3-4dcc-9030-46c8afb78842 req-e986b1fa-838b-4d58-8039-e9227e7ba1c6 service nova] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1331.174134] env[65107]: DEBUG oslo_concurrency.lockutils [req-1218d9e1-51d3-4dcc-9030-46c8afb78842 req-e986b1fa-838b-4d58-8039-e9227e7ba1c6 service nova] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1331.174254] env[65107]: DEBUG nova.compute.manager [req-1218d9e1-51d3-4dcc-9030-46c8afb78842 req-e986b1fa-838b-4d58-8039-e9227e7ba1c6 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] No waiting events found dispatching network-vif-plugged-a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1331.174419] env[65107]: WARNING nova.compute.manager [req-1218d9e1-51d3-4dcc-9030-46c8afb78842 req-e986b1fa-838b-4d58-8039-e9227e7ba1c6 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Received unexpected event network-vif-plugged-a900303a-2f4a-4b1f-b115-37e160d3834a for instance with vm_state building and task_state spawning. [ 1331.258036] env[65107]: DEBUG nova.network.neutron [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Successfully updated port: a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1331.760670] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1331.760859] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1331.761295] env[65107]: DEBUG nova.network.neutron [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1332.264159] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1332.264803] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1332.301930] env[65107]: DEBUG nova.network.neutron [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1332.322072] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1332.322578] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1332.389757] env[65107]: WARNING neutronclient.v2_0.client [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1332.390441] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1332.390788] env[65107]: WARNING openstack [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1332.477644] env[65107]: DEBUG nova.network.neutron [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating instance_info_cache with network_info: [{"id": "a900303a-2f4a-4b1f-b115-37e160d3834a", "address": "fa:16:3e:70:c4:0c", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa900303a-2f", "ovs_interfaceid": "a900303a-2f4a-4b1f-b115-37e160d3834a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1332.608286] env[65107]: INFO nova.compute.manager [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Swapping old allocation on dict_keys(['8adccf8d-6d73-43c0-ad15-ee89967ab8cd']) held by migration 2c5a4459-df77-4ef8-8722-39e760c95450 for instance [ 1332.633826] env[65107]: DEBUG nova.scheduler.client.report [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Overwriting current allocation {'allocations': {'8adccf8d-6d73-43c0-ad15-ee89967ab8cd': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 183}}, 'project_id': '139a38d3b7bd4d3a8861d2e1e6c7a2c1', 'user_id': '6723c29f05774e138ee11fa945c058e5', 'consumer_generation': 1} on consumer 84f88238-6b21-4530-8a08-72c4049f9150 {{(pid=65107) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1332.687532] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1332.718612] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.718838] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1332.719034] env[65107]: DEBUG nova.network.neutron [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1332.980769] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1332.981179] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Instance network_info: |[{"id": "a900303a-2f4a-4b1f-b115-37e160d3834a", "address": "fa:16:3e:70:c4:0c", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa900303a-2f", "ovs_interfaceid": "a900303a-2f4a-4b1f-b115-37e160d3834a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1332.981674] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:c4:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a900303a-2f4a-4b1f-b115-37e160d3834a', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1332.989776] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1332.990370] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1332.990611] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4652bba-5ccc-42fc-af01-29a981bdd7c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.010546] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1333.010546] env[65107]: value = "task-5103911" [ 1333.010546] env[65107]: _type = "Task" [ 1333.010546] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.019137] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103911, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.208012] env[65107]: DEBUG nova.compute.manager [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Received event network-changed-a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1333.208281] env[65107]: DEBUG nova.compute.manager [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Refreshing instance network info cache due to event network-changed-a900303a-2f4a-4b1f-b115-37e160d3834a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1333.208505] env[65107]: DEBUG oslo_concurrency.lockutils [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Acquiring lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1333.208651] env[65107]: DEBUG oslo_concurrency.lockutils [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Acquired lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1333.208868] env[65107]: DEBUG nova.network.neutron [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Refreshing network info cache for port a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1333.221840] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.222870] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1333.223429] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1333.348539] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1333.348934] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1333.413540] env[65107]: WARNING neutronclient.v2_0.client [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.414263] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1333.414603] env[65107]: WARNING openstack [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1333.494437] env[65107]: DEBUG nova.network.neutron [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [{"id": "ff8925f6-fab7-49a5-bae0-493ec802effc", "address": "fa:16:3e:14:0a:18", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff8925f6-fa", "ovs_interfaceid": "ff8925f6-fab7-49a5-bae0-493ec802effc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1333.521176] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103911, 'name': CreateVM_Task, 'duration_secs': 0.351702} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.521365] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1333.521852] env[65107]: WARNING neutronclient.v2_0.client [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.522309] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1333.522476] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1333.522798] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1333.523080] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f03deba-74a3-449b-b88f-a075f49943e0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.528659] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1333.528659] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52693987-772b-dff2-0ce2-1c078cb4aa50" [ 1333.528659] env[65107]: _type = "Task" [ 1333.528659] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.537231] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52693987-772b-dff2-0ce2-1c078cb4aa50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.711962] env[65107]: WARNING neutronclient.v2_0.client [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.712916] env[65107]: WARNING openstack [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1333.713141] env[65107]: WARNING openstack [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1333.805974] env[65107]: WARNING openstack [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1333.806422] env[65107]: WARNING openstack [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1333.868013] env[65107]: WARNING neutronclient.v2_0.client [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.868712] env[65107]: WARNING openstack [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1333.869086] env[65107]: WARNING openstack [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1333.955501] env[65107]: DEBUG nova.network.neutron [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updated VIF entry in instance network info cache for port a900303a-2f4a-4b1f-b115-37e160d3834a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1333.956069] env[65107]: DEBUG nova.network.neutron [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating instance_info_cache with network_info: [{"id": "a900303a-2f4a-4b1f-b115-37e160d3834a", "address": "fa:16:3e:70:c4:0c", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa900303a-2f", "ovs_interfaceid": "a900303a-2f4a-4b1f-b115-37e160d3834a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1333.998345] env[65107]: DEBUG oslo_concurrency.lockutils [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-84f88238-6b21-4530-8a08-72c4049f9150" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1333.998887] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1333.999130] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c63489a-d478-424b-9f97-ea1788c2b112 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.007032] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1334.007032] env[65107]: value = "task-5103912" [ 1334.007032] env[65107]: _type = "Task" [ 1334.007032] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.016273] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.042185] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52693987-772b-dff2-0ce2-1c078cb4aa50, 'name': SearchDatastore_Task, 'duration_secs': 0.0111} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.042458] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1334.042718] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1334.042975] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1334.043138] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1334.043322] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1334.043602] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db2d908e-b692-412d-a490-383d748e011a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.053674] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1334.053928] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1334.054785] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ecd090e-d8c3-40db-a81a-ed490ec3ab67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.060577] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1334.060577] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527d58c0-eb21-1651-4146-5ec801659f59" [ 1334.060577] env[65107]: _type = "Task" [ 1334.060577] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.069640] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527d58c0-eb21-1651-4146-5ec801659f59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.458946] env[65107]: DEBUG oslo_concurrency.lockutils [req-bd026db0-0236-4a0c-8662-dfd1926e490b req-5ce21466-82a2-4691-bfcd-835c61dd838c service nova] Releasing lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1334.517570] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103912, 'name': PowerOffVM_Task, 'duration_secs': 0.201391} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.517865] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1334.518598] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1334.518822] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1334.518983] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1334.519197] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1334.519352] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1334.519509] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1334.519732] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1334.519904] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1334.520090] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1334.520257] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1334.520433] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1334.525334] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4545515e-43bb-4571-82b2-c21b22ef22cf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.545240] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1334.545240] env[65107]: value = "task-5103913" [ 1334.545240] env[65107]: _type = "Task" [ 1334.545240] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.554187] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.570247] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527d58c0-eb21-1651-4146-5ec801659f59, 'name': SearchDatastore_Task, 'duration_secs': 0.009616} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.571069] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0435a90-5a92-4810-9275-f7c2985f59dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.577104] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1334.577104] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5218245b-04a2-cc92-fb23-a755b12559a3" [ 1334.577104] env[65107]: _type = "Task" [ 1334.577104] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.586886] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5218245b-04a2-cc92-fb23-a755b12559a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.056419] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103913, 'name': ReconfigVM_Task, 'duration_secs': 0.141276} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.057301] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38d63a1-afb0-442e-b090-80d91299a555 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.076838] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1335.077090] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1335.077255] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1335.077438] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1335.077587] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1335.077734] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1335.077943] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1335.078204] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1335.078399] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1335.078562] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1335.078737] env[65107]: DEBUG nova.virt.hardware [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1335.079618] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89e23359-f6a2-47e0-89bc-4323f876f75b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.088415] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1335.088415] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a0f4a-729a-593c-f6ed-b06f29450efc" [ 1335.088415] env[65107]: _type = "Task" [ 1335.088415] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.092137] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]5218245b-04a2-cc92-fb23-a755b12559a3, 'name': SearchDatastore_Task, 'duration_secs': 0.011734} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.095475] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1335.095782] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] f354b9e3-c426-48ca-8fe6-8bee94899ec8/f354b9e3-c426-48ca-8fe6-8bee94899ec8.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1335.096058] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39126d37-5d5d-4459-8cf4-b2cb6e1229c8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.104467] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]524a0f4a-729a-593c-f6ed-b06f29450efc, 'name': SearchDatastore_Task, 'duration_secs': 0.007612} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.110817] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfiguring VM instance instance-0000007b to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1335.111170] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1335.111170] env[65107]: value = "task-5103914" [ 1335.111170] env[65107]: _type = "Task" [ 1335.111170] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.111374] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b428a8ff-7c37-47a5-a22b-e7bfb811ab45 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.133906] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.135352] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1335.135352] env[65107]: value = "task-5103915" [ 1335.135352] env[65107]: _type = "Task" [ 1335.135352] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.143321] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103915, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.634238] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103914, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454957} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.634524] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] f354b9e3-c426-48ca-8fe6-8bee94899ec8/f354b9e3-c426-48ca-8fe6-8bee94899ec8.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1335.634768] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1335.635065] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e8378f0-1513-45f5-8ec4-aa26a3b0587b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.645300] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103915, 'name': ReconfigVM_Task, 'duration_secs': 0.196365} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.646612] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfigured VM instance instance-0000007b to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1335.647027] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1335.647027] env[65107]: value = "task-5103916" [ 1335.647027] env[65107]: _type = "Task" [ 1335.647027] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.647746] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a268da98-bec6-48c6-b06a-57a003ef62ba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.676009] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfiguring VM instance instance-0000007b to attach disk [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1335.676363] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.676619] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00483dca-fb3f-4048-b241-f2f19c6e3925 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.695923] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1335.695923] env[65107]: value = "task-5103917" [ 1335.695923] env[65107]: _type = "Task" [ 1335.695923] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1335.704682] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103917, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.160779] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071848} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.161053] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1336.161831] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc087ee7-f03f-4f67-88b7-51aa11acc995 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.183696] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfiguring VM instance instance-0000007d to attach disk [datastore1] f354b9e3-c426-48ca-8fe6-8bee94899ec8/f354b9e3-c426-48ca-8fe6-8bee94899ec8.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1336.184018] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb298fca-8d8f-40d1-9314-25c7bcc61c81 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.207675] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.209115] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1336.209115] env[65107]: value = "task-5103918" [ 1336.209115] env[65107]: _type = "Task" [ 1336.209115] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.217178] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.708566] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.718545] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103918, 'name': ReconfigVM_Task, 'duration_secs': 0.339606} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.718879] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfigured VM instance instance-0000007d to attach disk [datastore1] f354b9e3-c426-48ca-8fe6-8bee94899ec8/f354b9e3-c426-48ca-8fe6-8bee94899ec8.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1336.719597] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-918fd2e4-2050-44e8-98ee-e30ffbd4fda8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.727319] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1336.727319] env[65107]: value = "task-5103919" [ 1336.727319] env[65107]: _type = "Task" [ 1336.727319] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.735751] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103919, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.211235] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.241485] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103919, 'name': Rename_Task, 'duration_secs': 0.163377} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.241868] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1337.242209] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b95bd35-a1cb-46bf-bced-359b3b317284 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.251186] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1337.251186] env[65107]: value = "task-5103920" [ 1337.251186] env[65107]: _type = "Task" [ 1337.251186] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.260270] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.709356] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103917, 'name': ReconfigVM_Task, 'duration_secs': 1.875521} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.709772] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Reconfigured VM instance instance-0000007b to attach disk [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150/84f88238-6b21-4530-8a08-72c4049f9150.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1337.710596] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5da97f-1d43-41e6-a15b-1b3a43241eb5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.730362] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b71c16-a35f-4ab9-ac12-7330af0f396b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.749579] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20f2936-fd7c-45e3-ac02-77f3181081e9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.773844] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b255be45-5efd-401c-903a-c52172a2fbd3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.776645] env[65107]: DEBUG oslo_vmware.api [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103920, 'name': PowerOnVM_Task, 'duration_secs': 0.504444} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.776922] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1337.777150] env[65107]: INFO nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Took 6.87 seconds to spawn the instance on the hypervisor. [ 1337.777344] env[65107]: DEBUG nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1337.778452] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6067bd4a-b704-44bc-96d8-5c28de7fa734 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.783221] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1337.783742] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35803147-0f3f-4f23-8251-4805739b6180 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.791344] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1337.791344] env[65107]: value = "task-5103921" [ 1337.791344] env[65107]: _type = "Task" [ 1337.791344] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.803249] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.308800] env[65107]: INFO nova.compute.manager [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Took 11.61 seconds to build instance. [ 1338.314063] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103921, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.784558] env[65107]: DEBUG nova.compute.manager [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Received event network-changed-a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1338.784855] env[65107]: DEBUG nova.compute.manager [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Refreshing instance network info cache due to event network-changed-a900303a-2f4a-4b1f-b115-37e160d3834a. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1338.785034] env[65107]: DEBUG oslo_concurrency.lockutils [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Acquiring lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.785187] env[65107]: DEBUG oslo_concurrency.lockutils [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Acquired lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1338.785352] env[65107]: DEBUG nova.network.neutron [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Refreshing network info cache for port a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1338.806574] env[65107]: DEBUG oslo_vmware.api [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103921, 'name': PowerOnVM_Task, 'duration_secs': 0.648394} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.806842] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1338.810120] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.810346] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.810549] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.810730] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.810900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1338.812703] env[65107]: DEBUG oslo_concurrency.lockutils [None req-fdbd2d37-4379-4c1c-8c3a-26d0348e8530 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.119s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1338.813305] env[65107]: INFO nova.compute.manager [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Terminating instance [ 1339.288317] env[65107]: WARNING neutronclient.v2_0.client [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1339.289333] env[65107]: WARNING openstack [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1339.289451] env[65107]: WARNING openstack [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1339.317682] env[65107]: DEBUG nova.compute.manager [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1339.317954] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1339.318208] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-066b2f4f-9ec0-47b3-b76b-675945921b35 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.327734] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1339.327734] env[65107]: value = "task-5103922" [ 1339.327734] env[65107]: _type = "Task" [ 1339.327734] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.336879] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.434090] env[65107]: WARNING openstack [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1339.434571] env[65107]: WARNING openstack [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1339.505536] env[65107]: WARNING neutronclient.v2_0.client [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1339.506309] env[65107]: WARNING openstack [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1339.506684] env[65107]: WARNING openstack [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1339.590013] env[65107]: DEBUG nova.network.neutron [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updated VIF entry in instance network info cache for port a900303a-2f4a-4b1f-b115-37e160d3834a. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1339.590445] env[65107]: DEBUG nova.network.neutron [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating instance_info_cache with network_info: [{"id": "a900303a-2f4a-4b1f-b115-37e160d3834a", "address": "fa:16:3e:70:c4:0c", "network": {"id": "d7de159c-3d87-4da5-aa5e-8c41b6e6b70f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-730123040-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "84acb61dcc764cd79ad2e5b1e2850d16", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa900303a-2f", "ovs_interfaceid": "a900303a-2f4a-4b1f-b115-37e160d3834a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1339.823148] env[65107]: INFO nova.compute.manager [None req-7976b30a-5ab3-453f-8362-76bc2d4ff280 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance to original state: 'active' [ 1339.837682] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103922, 'name': PowerOffVM_Task, 'duration_secs': 0.195966} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.837932] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1339.838169] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1339.838380] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992912', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'name': 'volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6', 'attached_at': '2025-12-12T13:32:58.000000', 'detached_at': '', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'serial': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1339.839148] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5608930d-0f2b-4c7d-9856-f44e24df85d2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.858545] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9074900-9631-4e10-8ac0-8d82dcaf7606 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.866277] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19469414-d2ee-47c9-9a8d-ba152240fa4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.887097] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de57842-c55d-4a5c-84b4-28791df8bc93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.903124] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] The volume has not been displaced from its original location: [datastore2] volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93/volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1339.908338] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfiguring VM instance instance-0000007c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1339.908672] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe1d4c13-9938-4251-960d-554bbff6d515 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.929182] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1339.929182] env[65107]: value = "task-5103923" [ 1339.929182] env[65107]: _type = "Task" [ 1339.929182] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.937889] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103923, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.093826] env[65107]: DEBUG oslo_concurrency.lockutils [req-9c42918e-af85-4d7d-a82a-d828347b1ba8 req-e28d7fc5-c69c-4eeb-aab1-18e0756a1338 service nova] Releasing lock "refresh_cache-f354b9e3-c426-48ca-8fe6-8bee94899ec8" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1340.441052] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103923, 'name': ReconfigVM_Task, 'duration_secs': 0.178897} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.441270] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Reconfigured VM instance instance-0000007c to detach disk 2000 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1340.446068] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbcd1610-dd11-4720-9eb3-f08537a91b38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.464110] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1340.464110] env[65107]: value = "task-5103924" [ 1340.464110] env[65107]: _type = "Task" [ 1340.464110] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.474805] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103924, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.974907] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103924, 'name': ReconfigVM_Task, 'duration_secs': 0.154662} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.975606] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992912', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'name': 'volume-56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6', 'attached_at': '2025-12-12T13:32:58.000000', 'detached_at': '', 'volume_id': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93', 'serial': '56549d08-0e5b-4fd0-adad-3c3b55d4fb93'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1340.976055] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1340.976820] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6022c69-7213-4665-9cfc-c90772f121b8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.984143] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1340.984388] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f32c4cd3-534b-45bc-aac4-8c09d8ca7863 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.050989] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1341.051239] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1341.051427] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore2] f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1341.051717] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c726d4c-40b1-4721-805a-f4bfbbe4b700 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.058905] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1341.058905] env[65107]: value = "task-5103926" [ 1341.058905] env[65107]: _type = "Task" [ 1341.058905] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.067843] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103926, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.087468] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "84f88238-6b21-4530-8a08-72c4049f9150" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1341.087758] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1341.087988] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "84f88238-6b21-4530-8a08-72c4049f9150-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1341.088194] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1341.088370] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1341.090766] env[65107]: INFO nova.compute.manager [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Terminating instance [ 1341.569668] env[65107]: DEBUG oslo_vmware.api [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103926, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113516} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.569878] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1341.570056] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1341.570246] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1341.570421] env[65107]: INFO nova.compute.manager [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Took 2.25 seconds to destroy the instance on the hypervisor. [ 1341.570665] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1341.570856] env[65107]: DEBUG nova.compute.manager [-] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1341.570969] env[65107]: DEBUG nova.network.neutron [-] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1341.571225] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1341.571741] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1341.571996] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1341.595013] env[65107]: DEBUG nova.compute.manager [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1341.595236] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1341.596220] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfccc6ad-1cb8-4272-b2c9-cce7149b0558 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.605762] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1341.606060] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05aaf577-3794-4899-b16c-4b893d253085 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.612748] env[65107]: DEBUG oslo_vmware.api [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1341.612748] env[65107]: value = "task-5103927" [ 1341.612748] env[65107]: _type = "Task" [ 1341.612748] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.617793] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1341.623651] env[65107]: DEBUG oslo_vmware.api [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.069806] env[65107]: DEBUG nova.compute.manager [req-220497d0-56a6-4e93-9b91-51de75bd7234 req-b435cec2-acad-4dfe-85bd-a279b68229fc service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Received event network-vif-deleted-30bb95a2-1900-433c-98bc-b8436565a0ff {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1342.070178] env[65107]: INFO nova.compute.manager [req-220497d0-56a6-4e93-9b91-51de75bd7234 req-b435cec2-acad-4dfe-85bd-a279b68229fc service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Neutron deleted interface 30bb95a2-1900-433c-98bc-b8436565a0ff; detaching it from the instance and deleting it from the info cache [ 1342.070318] env[65107]: DEBUG nova.network.neutron [req-220497d0-56a6-4e93-9b91-51de75bd7234 req-b435cec2-acad-4dfe-85bd-a279b68229fc service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1342.123495] env[65107]: DEBUG oslo_vmware.api [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103927, 'name': PowerOffVM_Task, 'duration_secs': 0.218848} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.123760] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1342.123933] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1342.124193] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70f7ad21-fdef-49de-9c15-6a26f9d20366 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.193604] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1342.193604] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1342.193929] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleting the datastore file [datastore2] 84f88238-6b21-4530-8a08-72c4049f9150 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1342.194045] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1618df23-ec9b-45d9-a1b1-ae8c92b78ee4 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.204272] env[65107]: DEBUG oslo_vmware.api [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1342.204272] env[65107]: value = "task-5103929" [ 1342.204272] env[65107]: _type = "Task" [ 1342.204272] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.216215] env[65107]: DEBUG oslo_vmware.api [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.547171] env[65107]: DEBUG nova.network.neutron [-] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1342.573215] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42c42014-8fbb-43ea-a3e9-2e000bb36158 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.583399] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b020f138-f59e-4d19-8b73-de2091ac4464 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.611919] env[65107]: DEBUG nova.compute.manager [req-220497d0-56a6-4e93-9b91-51de75bd7234 req-b435cec2-acad-4dfe-85bd-a279b68229fc service nova] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Detach interface failed, port_id=30bb95a2-1900-433c-98bc-b8436565a0ff, reason: Instance f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1342.714708] env[65107]: DEBUG oslo_vmware.api [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20395} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.714995] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1342.715141] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1342.715333] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1342.715499] env[65107]: INFO nova.compute.manager [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1342.715748] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1342.716037] env[65107]: DEBUG nova.compute.manager [-] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1342.716156] env[65107]: DEBUG nova.network.neutron [-] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1342.716402] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1342.717055] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1342.717232] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1342.759477] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1343.050659] env[65107]: INFO nova.compute.manager [-] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Took 1.48 seconds to deallocate network for instance. [ 1343.597547] env[65107]: INFO nova.compute.manager [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Took 0.55 seconds to detach 1 volumes for instance. [ 1343.599771] env[65107]: DEBUG nova.compute.manager [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Deleting volume: 56549d08-0e5b-4fd0-adad-3c3b55d4fb93 {{(pid=65107) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1343.650773] env[65107]: DEBUG nova.network.neutron [-] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1344.096983] env[65107]: DEBUG nova.compute.manager [req-c29cc8de-10a9-40a5-a446-166e30edf8f3 req-31db33f8-eef7-4bd3-bd7e-3009fce39b1d service nova] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Received event network-vif-deleted-ff8925f6-fab7-49a5-bae0-493ec802effc {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1344.139041] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1344.139235] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1344.139455] env[65107]: DEBUG nova.objects.instance [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'resources' on Instance uuid f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1344.153787] env[65107]: INFO nova.compute.manager [-] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Took 1.44 seconds to deallocate network for instance. [ 1344.660181] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1344.722601] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551a6f5b-8d48-4174-8dbb-22bf8c27698d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.730912] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8bb777-9635-434d-91f6-74090bf880ed {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.764020] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9e88ec-445a-4a39-8d8a-532102805b10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.772802] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d731d84-3db9-4667-986c-a9be3ec1270c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.788070] env[65107]: DEBUG nova.compute.provider_tree [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1345.292085] env[65107]: DEBUG nova.scheduler.client.report [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1345.797650] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1345.800271] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.140s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1345.800477] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1345.819227] env[65107]: INFO nova.scheduler.client.report [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted allocations for instance f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6 [ 1345.825630] env[65107]: INFO nova.scheduler.client.report [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted allocations for instance 84f88238-6b21-4530-8a08-72c4049f9150 [ 1346.331615] env[65107]: DEBUG oslo_concurrency.lockutils [None req-231fd39a-4833-4e38-9fd0-aa681ab36ed1 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.521s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1346.332693] env[65107]: DEBUG oslo_concurrency.lockutils [None req-12d603c1-cc7d-4402-bb93-8c28598ab4c5 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "84f88238-6b21-4530-8a08-72c4049f9150" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.245s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1346.628965] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1346.629283] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1346.629508] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1346.630224] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1346.630413] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1346.632775] env[65107]: INFO nova.compute.manager [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Terminating instance [ 1347.138994] env[65107]: DEBUG nova.compute.manager [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1347.139403] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1347.139872] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "995c579b-9e18-4d3d-816d-0b48f352f938" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1347.140095] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1347.141743] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a89e778-2d66-4dba-8da0-430d5257a6d9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.151198] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1347.152056] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-365fb505-133e-423f-b38b-10e481256719 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.159136] env[65107]: DEBUG oslo_vmware.api [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1347.159136] env[65107]: value = "task-5103931" [ 1347.159136] env[65107]: _type = "Task" [ 1347.159136] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.169833] env[65107]: DEBUG oslo_vmware.api [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.646056] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1347.669730] env[65107]: DEBUG oslo_vmware.api [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103931, 'name': PowerOffVM_Task, 'duration_secs': 0.235522} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.670028] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1347.670217] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1347.670479] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af1bcc5e-a5fb-4155-9fb6-5dc378cc0c24 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.740078] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1347.740314] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1347.740556] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore1] e72d0cb3-027c-49af-aa23-b791ae2f5fcf {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1347.740893] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d4e4d66-6d0e-46b1-b1a5-42c506c96aa9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.748742] env[65107]: DEBUG oslo_vmware.api [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1347.748742] env[65107]: value = "task-5103933" [ 1347.748742] env[65107]: _type = "Task" [ 1347.748742] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.757172] env[65107]: DEBUG oslo_vmware.api [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.166867] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1348.167195] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1348.168792] env[65107]: INFO nova.compute.claims [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1348.259095] env[65107]: DEBUG oslo_vmware.api [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135396} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.259376] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1348.259562] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1348.259753] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1348.260052] env[65107]: INFO nova.compute.manager [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1348.260418] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1348.260686] env[65107]: DEBUG nova.compute.manager [-] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1348.260821] env[65107]: DEBUG nova.network.neutron [-] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1348.261138] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1348.261828] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1348.262116] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1348.299054] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1348.544777] env[65107]: DEBUG nova.compute.manager [req-6d8d843a-b7d2-40b3-bf29-908a26fa068f req-8356f6c1-721c-4acd-bd2d-64ae1097fb27 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Received event network-vif-deleted-d4ef2ece-8832-4198-a869-6869223cf87a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1348.544936] env[65107]: INFO nova.compute.manager [req-6d8d843a-b7d2-40b3-bf29-908a26fa068f req-8356f6c1-721c-4acd-bd2d-64ae1097fb27 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Neutron deleted interface d4ef2ece-8832-4198-a869-6869223cf87a; detaching it from the instance and deleting it from the info cache [ 1348.545049] env[65107]: DEBUG nova.network.neutron [req-6d8d843a-b7d2-40b3-bf29-908a26fa068f req-8356f6c1-721c-4acd-bd2d-64ae1097fb27 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1349.026686] env[65107]: DEBUG nova.network.neutron [-] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1349.047596] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3431332-9a31-458f-8ccb-578f83580692 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.057978] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5d2057-f10d-4f05-82e7-27ff40d4c7c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.086192] env[65107]: DEBUG nova.compute.manager [req-6d8d843a-b7d2-40b3-bf29-908a26fa068f req-8356f6c1-721c-4acd-bd2d-64ae1097fb27 service nova] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Detach interface failed, port_id=d4ef2ece-8832-4198-a869-6869223cf87a, reason: Instance e72d0cb3-027c-49af-aa23-b791ae2f5fcf could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1349.237640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7603114e-66ca-4aeb-a40d-00741c443787 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.245864] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48051f1c-bd1f-4732-bd4d-d2081bc07537 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.279247] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5630665f-2d87-4e96-b89a-8bedb69f710b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.287745] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cace5777-2f13-44fb-9f4b-ffaac2c72f59 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.303249] env[65107]: DEBUG nova.compute.provider_tree [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1349.530137] env[65107]: INFO nova.compute.manager [-] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Took 1.27 seconds to deallocate network for instance. [ 1349.806294] env[65107]: DEBUG nova.scheduler.client.report [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1350.036982] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1350.310947] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1350.311510] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1350.314276] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.277s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1350.315115] env[65107]: DEBUG nova.objects.instance [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'resources' on Instance uuid e72d0cb3-027c-49af-aa23-b791ae2f5fcf {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1350.818159] env[65107]: DEBUG nova.compute.utils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1350.823028] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1350.823179] env[65107]: DEBUG nova.network.neutron [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1350.823478] env[65107]: WARNING neutronclient.v2_0.client [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1350.823788] env[65107]: WARNING neutronclient.v2_0.client [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1350.824385] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1350.824728] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1350.878059] env[65107]: DEBUG nova.policy [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6723c29f05774e138ee11fa945c058e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '139a38d3b7bd4d3a8861d2e1e6c7a2c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1350.892544] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ea37f6-c0c3-4a5c-b091-c2457072620f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.900635] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa20221c-59a6-4d76-9f2b-622c87c247dd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.935599] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c47e4fc-6b49-461c-987f-be5df1d8f849 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.944313] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00511902-cc75-433c-8f3f-ecc4c12517fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.959118] env[65107]: DEBUG nova.compute.provider_tree [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1351.216598] env[65107]: DEBUG nova.network.neutron [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Successfully created port: 187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1351.324330] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1351.462158] env[65107]: DEBUG nova.scheduler.client.report [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1351.967065] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.653s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1351.986517] env[65107]: INFO nova.scheduler.client.report [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted allocations for instance e72d0cb3-027c-49af-aa23-b791ae2f5fcf [ 1352.335435] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1352.361738] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1352.362013] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1352.362255] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1352.362484] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1352.362620] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1352.362769] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1352.362984] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1352.363162] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1352.363332] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1352.363494] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1352.363667] env[65107]: DEBUG nova.virt.hardware [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1352.364635] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd275bd-511e-48df-906c-3785434d61dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.374332] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce21251-dad6-4b1b-8d5a-654ab8720bdf {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.497230] env[65107]: DEBUG oslo_concurrency.lockutils [None req-8b7cc9ee-f813-415e-946e-40c6f4941479 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "e72d0cb3-027c-49af-aa23-b791ae2f5fcf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.868s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1352.619423] env[65107]: DEBUG nova.compute.manager [req-fae1a27e-2f6a-46c0-9ded-72a30eb56ccd req-07f37bbc-7dc6-4cfc-a162-fc4a15e7122a service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Received event network-vif-plugged-187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1352.619586] env[65107]: DEBUG oslo_concurrency.lockutils [req-fae1a27e-2f6a-46c0-9ded-72a30eb56ccd req-07f37bbc-7dc6-4cfc-a162-fc4a15e7122a service nova] Acquiring lock "995c579b-9e18-4d3d-816d-0b48f352f938-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1352.619805] env[65107]: DEBUG oslo_concurrency.lockutils [req-fae1a27e-2f6a-46c0-9ded-72a30eb56ccd req-07f37bbc-7dc6-4cfc-a162-fc4a15e7122a service nova] Lock "995c579b-9e18-4d3d-816d-0b48f352f938-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1352.619976] env[65107]: DEBUG oslo_concurrency.lockutils [req-fae1a27e-2f6a-46c0-9ded-72a30eb56ccd req-07f37bbc-7dc6-4cfc-a162-fc4a15e7122a service nova] Lock "995c579b-9e18-4d3d-816d-0b48f352f938-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1352.620161] env[65107]: DEBUG nova.compute.manager [req-fae1a27e-2f6a-46c0-9ded-72a30eb56ccd req-07f37bbc-7dc6-4cfc-a162-fc4a15e7122a service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] No waiting events found dispatching network-vif-plugged-187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1352.620353] env[65107]: WARNING nova.compute.manager [req-fae1a27e-2f6a-46c0-9ded-72a30eb56ccd req-07f37bbc-7dc6-4cfc-a162-fc4a15e7122a service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Received unexpected event network-vif-plugged-187b4d9a-9a89-4e19-8674-5bd91faf1d78 for instance with vm_state building and task_state spawning. [ 1352.721961] env[65107]: DEBUG nova.network.neutron [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Successfully updated port: 187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1353.224528] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1353.224715] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1353.224898] env[65107]: DEBUG nova.network.neutron [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1353.728428] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1353.728939] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.765141] env[65107]: DEBUG nova.network.neutron [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1353.787711] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1353.788146] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.852530] env[65107]: WARNING neutronclient.v2_0.client [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1353.853205] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1353.853548] env[65107]: WARNING openstack [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.940187] env[65107]: DEBUG nova.network.neutron [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [{"id": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "address": "fa:16:3e:9e:0c:e8", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap187b4d9a-9a", "ovs_interfaceid": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1354.444380] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1354.444782] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Instance network_info: |[{"id": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "address": "fa:16:3e:9e:0c:e8", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap187b4d9a-9a", "ovs_interfaceid": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1354.445320] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:0c:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f9ffee1-f413-4f28-8bc4-3fb2cf299789', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '187b4d9a-9a89-4e19-8674-5bd91faf1d78', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1354.452764] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1354.452991] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1354.453280] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b0afe09-3028-4ef9-9c4a-f441303e0e7e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.473726] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1354.473726] env[65107]: value = "task-5103935" [ 1354.473726] env[65107]: _type = "Task" [ 1354.473726] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.482086] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103935, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.647574] env[65107]: DEBUG nova.compute.manager [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Received event network-changed-187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1354.647682] env[65107]: DEBUG nova.compute.manager [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Refreshing instance network info cache due to event network-changed-187b4d9a-9a89-4e19-8674-5bd91faf1d78. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1354.647902] env[65107]: DEBUG oslo_concurrency.lockutils [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Acquiring lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1354.648248] env[65107]: DEBUG oslo_concurrency.lockutils [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Acquired lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1354.648480] env[65107]: DEBUG nova.network.neutron [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Refreshing network info cache for port 187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1354.984321] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103935, 'name': CreateVM_Task, 'duration_secs': 0.317993} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.984702] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1354.985092] env[65107]: WARNING neutronclient.v2_0.client [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1354.985581] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1354.985773] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1354.986159] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1354.986465] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62c927f2-fc18-4253-87a4-a403cf1de84c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.991530] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1354.991530] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527be289-2b50-51b3-0c0c-85ef33839597" [ 1354.991530] env[65107]: _type = "Task" [ 1354.991530] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.999874] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527be289-2b50-51b3-0c0c-85ef33839597, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.150871] env[65107]: WARNING neutronclient.v2_0.client [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1355.151686] env[65107]: WARNING openstack [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1355.152081] env[65107]: WARNING openstack [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1355.258031] env[65107]: WARNING openstack [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1355.258668] env[65107]: WARNING openstack [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1355.321266] env[65107]: WARNING neutronclient.v2_0.client [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1355.322093] env[65107]: WARNING openstack [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1355.322355] env[65107]: WARNING openstack [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1355.404315] env[65107]: DEBUG nova.network.neutron [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updated VIF entry in instance network info cache for port 187b4d9a-9a89-4e19-8674-5bd91faf1d78. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1355.404705] env[65107]: DEBUG nova.network.neutron [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [{"id": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "address": "fa:16:3e:9e:0c:e8", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap187b4d9a-9a", "ovs_interfaceid": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1355.501876] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]527be289-2b50-51b3-0c0c-85ef33839597, 'name': SearchDatastore_Task, 'duration_secs': 0.010935} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.502146] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1355.502397] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1355.502636] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1355.502781] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1355.502957] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1355.503250] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-738998ea-fefb-4450-addf-9d9c097fd73e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.512563] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1355.512755] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1355.513567] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-643669cd-e4a2-44a6-aa87-59dc88bfa638 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.519701] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1355.519701] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee7c89-6c49-8cb2-4960-35c714f9d369" [ 1355.519701] env[65107]: _type = "Task" [ 1355.519701] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.528175] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee7c89-6c49-8cb2-4960-35c714f9d369, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.907600] env[65107]: DEBUG oslo_concurrency.lockutils [req-cc6a5bb7-0d8f-4e9e-afe9-5f8296b44313 req-76d1a809-9576-4d12-8ff9-878807523210 service nova] Releasing lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1356.030865] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52ee7c89-6c49-8cb2-4960-35c714f9d369, 'name': SearchDatastore_Task, 'duration_secs': 0.0114} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.031775] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ebc231c-393f-404f-ba58-e113f62c4c53 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.038228] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1356.038228] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b2eb9b-a263-021b-eb86-51edc099f46f" [ 1356.038228] env[65107]: _type = "Task" [ 1356.038228] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.049156] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b2eb9b-a263-021b-eb86-51edc099f46f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.405208] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "3df65e64-0c45-4707-960e-8f1767e2d011" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1356.405511] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "3df65e64-0c45-4707-960e-8f1767e2d011" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1356.405730] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "3df65e64-0c45-4707-960e-8f1767e2d011-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1356.405911] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "3df65e64-0c45-4707-960e-8f1767e2d011-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1356.406143] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "3df65e64-0c45-4707-960e-8f1767e2d011-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1356.408477] env[65107]: INFO nova.compute.manager [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Terminating instance [ 1356.548927] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52b2eb9b-a263-021b-eb86-51edc099f46f, 'name': SearchDatastore_Task, 'duration_secs': 0.010824} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.549530] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1356.549801] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 995c579b-9e18-4d3d-816d-0b48f352f938/995c579b-9e18-4d3d-816d-0b48f352f938.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1356.550074] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a22bfaa-196b-4b89-b2eb-ac6fff6f5af8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.557130] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1356.557130] env[65107]: value = "task-5103936" [ 1356.557130] env[65107]: _type = "Task" [ 1356.557130] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.565889] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103936, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.912925] env[65107]: DEBUG nova.compute.manager [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1356.913252] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1356.914274] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c47416-d242-4605-b94d-d9f3831ede04 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.923152] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1356.923524] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78fe7fa8-7f41-4b86-b69a-175f32ae94f7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.933630] env[65107]: DEBUG oslo_vmware.api [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1356.933630] env[65107]: value = "task-5103937" [ 1356.933630] env[65107]: _type = "Task" [ 1356.933630] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.943984] env[65107]: DEBUG oslo_vmware.api [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.068370] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103936, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495432} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.068700] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore2] 995c579b-9e18-4d3d-816d-0b48f352f938/995c579b-9e18-4d3d-816d-0b48f352f938.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1357.068871] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1357.069186] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f227a6e7-59ad-4dec-b975-705e92e51dea {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.077384] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1357.077384] env[65107]: value = "task-5103938" [ 1357.077384] env[65107]: _type = "Task" [ 1357.077384] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.090973] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.443713] env[65107]: DEBUG oslo_vmware.api [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103937, 'name': PowerOffVM_Task, 'duration_secs': 0.278078} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.443999] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1357.444207] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1357.444468] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba3984e3-810a-4282-9bb0-30635ede494a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.509749] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1357.510050] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1357.510201] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleting the datastore file [datastore1] 3df65e64-0c45-4707-960e-8f1767e2d011 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1357.510486] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f72b4cca-6a4a-4d8e-a0d9-4d074cf43285 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.517244] env[65107]: DEBUG oslo_vmware.api [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for the task: (returnval){ [ 1357.517244] env[65107]: value = "task-5103940" [ 1357.517244] env[65107]: _type = "Task" [ 1357.517244] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.525370] env[65107]: DEBUG oslo_vmware.api [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103940, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.588611] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069142} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.588828] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1357.589691] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f325a24b-3a8a-472e-8fcd-4d793664f65a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.612340] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Reconfiguring VM instance instance-0000007e to attach disk [datastore2] 995c579b-9e18-4d3d-816d-0b48f352f938/995c579b-9e18-4d3d-816d-0b48f352f938.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1357.612649] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20ce530e-83d4-444f-b24d-398cfedecc87 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.633710] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1357.633710] env[65107]: value = "task-5103941" [ 1357.633710] env[65107]: _type = "Task" [ 1357.633710] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.642451] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103941, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.029292] env[65107]: DEBUG oslo_vmware.api [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Task: {'id': task-5103940, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154214} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.029604] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1358.029816] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1358.030030] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1358.030274] env[65107]: INFO nova.compute.manager [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1358.030579] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1358.030855] env[65107]: DEBUG nova.compute.manager [-] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1358.030960] env[65107]: DEBUG nova.network.neutron [-] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1358.031282] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1358.031872] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1358.032198] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1358.071182] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1358.146535] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103941, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.342625] env[65107]: DEBUG nova.compute.manager [req-48f90261-eb2b-4c68-804a-d854fa69f087 req-fa48bc4b-7195-4a82-a0a2-aaf2c2d43c16 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Received event network-vif-deleted-75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1358.342798] env[65107]: INFO nova.compute.manager [req-48f90261-eb2b-4c68-804a-d854fa69f087 req-fa48bc4b-7195-4a82-a0a2-aaf2c2d43c16 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Neutron deleted interface 75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d; detaching it from the instance and deleting it from the info cache [ 1358.343018] env[65107]: DEBUG nova.network.neutron [req-48f90261-eb2b-4c68-804a-d854fa69f087 req-fa48bc4b-7195-4a82-a0a2-aaf2c2d43c16 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1358.647496] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103941, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.817574] env[65107]: DEBUG nova.network.neutron [-] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1358.845784] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c3d7fd9-77a7-435d-9eac-c0bcb610b8c3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.857207] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336a2ad6-57c7-4e2e-bd52-81140fcefd70 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.890665] env[65107]: DEBUG nova.compute.manager [req-48f90261-eb2b-4c68-804a-d854fa69f087 req-fa48bc4b-7195-4a82-a0a2-aaf2c2d43c16 service nova] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Detach interface failed, port_id=75cfc6b6-0af5-41f1-9ccf-8f5a86fdfd1d, reason: Instance 3df65e64-0c45-4707-960e-8f1767e2d011 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1359.149785] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103941, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.321743] env[65107]: INFO nova.compute.manager [-] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Took 1.29 seconds to deallocate network for instance. [ 1359.647513] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103941, 'name': ReconfigVM_Task, 'duration_secs': 1.591479} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.647812] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Reconfigured VM instance instance-0000007e to attach disk [datastore2] 995c579b-9e18-4d3d-816d-0b48f352f938/995c579b-9e18-4d3d-816d-0b48f352f938.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1359.648512] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38906c2b-de0e-404f-909a-2b11763f0bf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.656523] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1359.656523] env[65107]: value = "task-5103942" [ 1359.656523] env[65107]: _type = "Task" [ 1359.656523] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.665884] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103942, 'name': Rename_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.828892] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1359.829157] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1359.829404] env[65107]: DEBUG nova.objects.instance [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lazy-loading 'resources' on Instance uuid 3df65e64-0c45-4707-960e-8f1767e2d011 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1360.167749] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103942, 'name': Rename_Task, 'duration_secs': 0.164781} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.168222] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1360.168222] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ddfc9998-f9b9-4e7d-ba1a-28b1a69a9e1e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.175872] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1360.175872] env[65107]: value = "task-5103943" [ 1360.175872] env[65107]: _type = "Task" [ 1360.175872] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.184437] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103943, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.351984] env[65107]: DEBUG nova.scheduler.client.report [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Refreshing inventories for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1360.366967] env[65107]: DEBUG nova.scheduler.client.report [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Updating ProviderTree inventory for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1360.367275] env[65107]: DEBUG nova.compute.provider_tree [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Updating inventory in ProviderTree for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1360.379822] env[65107]: DEBUG nova.scheduler.client.report [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Refreshing aggregate associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, aggregates: None {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1360.399016] env[65107]: DEBUG nova.scheduler.client.report [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Refreshing trait associations for resource provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd, traits: COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=65107) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1360.453324] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fdac59-bc5f-4777-8f17-dbdf615c6518 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.461046] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083ea540-5dc6-41d1-a1f2-815bb945d3a7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.492329] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738c31a8-5769-4d22-97bc-a2c3dbea3303 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.500435] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d98fcc-250d-4d03-ae02-525a15e2ab1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.514301] env[65107]: DEBUG nova.compute.provider_tree [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1360.686668] env[65107]: DEBUG oslo_vmware.api [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103943, 'name': PowerOnVM_Task, 'duration_secs': 0.466345} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.686925] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1360.687156] env[65107]: INFO nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Took 8.35 seconds to spawn the instance on the hypervisor. [ 1360.687350] env[65107]: DEBUG nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1360.688152] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6572872-92ed-4e69-a80d-9168f383fa7b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.017509] env[65107]: DEBUG nova.scheduler.client.report [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1361.205757] env[65107]: INFO nova.compute.manager [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Took 13.06 seconds to build instance. [ 1361.360746] env[65107]: DEBUG nova.compute.manager [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Received event network-changed-187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1361.360973] env[65107]: DEBUG nova.compute.manager [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Refreshing instance network info cache due to event network-changed-187b4d9a-9a89-4e19-8674-5bd91faf1d78. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1361.361262] env[65107]: DEBUG oslo_concurrency.lockutils [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Acquiring lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.361495] env[65107]: DEBUG oslo_concurrency.lockutils [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Acquired lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1361.361726] env[65107]: DEBUG nova.network.neutron [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Refreshing network info cache for port 187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1361.522282] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1361.545294] env[65107]: INFO nova.scheduler.client.report [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Deleted allocations for instance 3df65e64-0c45-4707-960e-8f1767e2d011 [ 1361.708188] env[65107]: DEBUG oslo_concurrency.lockutils [None req-57f214f9-a196-48f3-8bb7-fe97dad5156c tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.568s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1361.864495] env[65107]: WARNING neutronclient.v2_0.client [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1361.865461] env[65107]: WARNING openstack [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1361.865923] env[65107]: WARNING openstack [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1361.979320] env[65107]: WARNING openstack [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1361.979725] env[65107]: WARNING openstack [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1362.040690] env[65107]: WARNING neutronclient.v2_0.client [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1362.041376] env[65107]: WARNING openstack [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1362.041725] env[65107]: WARNING openstack [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1362.052192] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a6c6c3d6-1163-4064-8a58-c2a2c2cc3772 tempest-ServerActionsTestOtherA-567098223 tempest-ServerActionsTestOtherA-567098223-project-member] Lock "3df65e64-0c45-4707-960e-8f1767e2d011" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.647s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1362.126243] env[65107]: DEBUG nova.network.neutron [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updated VIF entry in instance network info cache for port 187b4d9a-9a89-4e19-8674-5bd91faf1d78. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1362.126688] env[65107]: DEBUG nova.network.neutron [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [{"id": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "address": "fa:16:3e:9e:0c:e8", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap187b4d9a-9a", "ovs_interfaceid": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1362.630288] env[65107]: DEBUG oslo_concurrency.lockutils [req-ab331180-fbd9-4fdc-9cb8-67dc5a5903a2 req-7fbb723f-1cac-453b-9742-6aa2559e0181 service nova] Releasing lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1375.857384] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1375.857719] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1376.248202] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1376.248463] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1376.248579] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1376.361202] env[65107]: DEBUG nova.compute.utils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1376.496012] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1376.864454] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1377.499735] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1377.927049] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1377.927487] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1377.927774] env[65107]: INFO nova.compute.manager [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Attaching volume 4bde42ea-1ce0-45f7-9ea7-17c50ae3db26 to /dev/sdb [ 1377.957953] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da5b035-7657-4ef7-92d4-dbb6847f3f1b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.966865] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6829afe7-5536-4dd6-b82d-695a1050fd74 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.980207] env[65107]: DEBUG nova.virt.block_device [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating existing volume attachment record: 1599029b-2cee-4af3-9227-b7042724124a {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1378.500895] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.499632] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.499933] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.523529] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1382.523897] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992917', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'name': 'volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'serial': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1382.524903] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa646a1-a295-41a1-8168-2c1934cd71f2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.542721] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8805868-a30f-4800-9760-7b3d20f2b2fb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.567292] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfiguring VM instance instance-0000007d to attach disk [datastore1] volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26/volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1382.567558] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-376c07c3-0559-442e-8f53-78db9c510b69 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.586689] env[65107]: DEBUG oslo_vmware.api [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1382.586689] env[65107]: value = "task-5103946" [ 1382.586689] env[65107]: _type = "Task" [ 1382.586689] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.594953] env[65107]: DEBUG oslo_vmware.api [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.097408] env[65107]: DEBUG oslo_vmware.api [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103946, 'name': ReconfigVM_Task, 'duration_secs': 0.334922} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1383.097672] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfigured VM instance instance-0000007d to attach disk [datastore1] volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26/volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1383.102263] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ccb76cf-7790-4737-b598-aeef1898b0c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.117938] env[65107]: DEBUG oslo_vmware.api [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1383.117938] env[65107]: value = "task-5103947" [ 1383.117938] env[65107]: _type = "Task" [ 1383.117938] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1383.126318] env[65107]: DEBUG oslo_vmware.api [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103947, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1383.627568] env[65107]: DEBUG oslo_vmware.api [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103947, 'name': ReconfigVM_Task, 'duration_secs': 0.134496} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1383.627945] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992917', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'name': 'volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'serial': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1384.500184] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.662659] env[65107]: DEBUG nova.objects.instance [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid f354b9e3-c426-48ca-8fe6-8bee94899ec8 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1385.002973] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1385.003588] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1385.003588] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1385.003785] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1385.004640] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888a1280-8adf-47df-8b49-8e1db8ec7905 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.015010] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78764d1f-f639-49c0-bb54-3c0cf9fe3b6e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.029417] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8afc89-0988-4c0b-add1-9f0c255d452f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.036096] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf13f94-17e2-4106-a537-73f592a3a152 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.066169] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180228MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1385.066333] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1385.066664] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1385.168755] env[65107]: DEBUG oslo_concurrency.lockutils [None req-3625f41f-f58f-4b31-bbdc-aafe30b83aaa tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1386.003257] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1386.003585] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1386.091747] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance f354b9e3-c426-48ca-8fe6-8bee94899ec8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1386.091922] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Instance 995c579b-9e18-4d3d-816d-0b48f352f938 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65107) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1386.092127] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1386.092278] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=100GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '2', 'num_vm_active': '2', 'num_task_None': '2', 'num_os_type_None': '2', 'num_proj_84acb61dcc764cd79ad2e5b1e2850d16': '1', 'io_workload': '0', 'num_proj_139a38d3b7bd4d3a8861d2e1e6c7a2c1': '1'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1386.130774] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b528dbb-bad7-4e07-82d3-4d13b536cf93 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.138463] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d01382-7ed4-4b5f-be59-f51e4aaf0db3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.170221] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c4a19b-d4ef-4ae5-93ab-dc50a79cc898 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.177638] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4673bd-df2e-4f7e-aa5b-3f17c3a61730 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.191795] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1386.507034] env[65107]: DEBUG nova.compute.utils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1386.694907] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1387.010450] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1387.200359] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1387.200581] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.134s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1388.068618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1388.069049] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1388.069049] env[65107]: INFO nova.compute.manager [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Attaching volume bc475fc5-e5ae-48d2-97c4-33913a95981e to /dev/sdc [ 1388.100580] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01202ed5-33ca-4a40-af69-eb3946d71921 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.108385] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b55bbf-4195-4b44-8cc5-32431c23d09a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.124030] env[65107]: DEBUG nova.virt.block_device [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating existing volume attachment record: 9bc97d96-562b-4250-a052-ae97542c1297 {{(pid=65107) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1391.195336] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.671601] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Volume attach. Driver type: vmdk {{(pid=65107) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1392.672090] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992918', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'name': 'volume-bc475fc5-e5ae-48d2-97c4-33913a95981e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'serial': 'bc475fc5-e5ae-48d2-97c4-33913a95981e'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1392.673151] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b265d544-61b2-44d3-8470-4aa316e9ffa7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.690749] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f924867a-c356-43e8-8e31-f873f75bbe39 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.718672] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfiguring VM instance instance-0000007d to attach disk [datastore1] volume-bc475fc5-e5ae-48d2-97c4-33913a95981e/volume-bc475fc5-e5ae-48d2-97c4-33913a95981e.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1392.719032] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8196a3e0-7848-4e6e-97f0-6be928719f77 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.738388] env[65107]: DEBUG oslo_vmware.api [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1392.738388] env[65107]: value = "task-5103950" [ 1392.738388] env[65107]: _type = "Task" [ 1392.738388] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.746830] env[65107]: DEBUG oslo_vmware.api [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103950, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.248570] env[65107]: DEBUG oslo_vmware.api [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103950, 'name': ReconfigVM_Task, 'duration_secs': 0.36946} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.248883] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfigured VM instance instance-0000007d to attach disk [datastore1] volume-bc475fc5-e5ae-48d2-97c4-33913a95981e/volume-bc475fc5-e5ae-48d2-97c4-33913a95981e.vmdk or device None with type thin {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1393.253529] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6341c355-5289-487f-b1dc-101d486ded8c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.269721] env[65107]: DEBUG oslo_vmware.api [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1393.269721] env[65107]: value = "task-5103951" [ 1393.269721] env[65107]: _type = "Task" [ 1393.269721] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.279935] env[65107]: DEBUG oslo_vmware.api [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103951, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.780361] env[65107]: DEBUG oslo_vmware.api [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103951, 'name': ReconfigVM_Task, 'duration_secs': 0.148522} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.780680] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992918', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'name': 'volume-bc475fc5-e5ae-48d2-97c4-33913a95981e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'serial': 'bc475fc5-e5ae-48d2-97c4-33913a95981e'} {{(pid=65107) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1394.822708] env[65107]: DEBUG nova.objects.instance [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid f354b9e3-c426-48ca-8fe6-8bee94899ec8 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1395.328216] env[65107]: DEBUG oslo_concurrency.lockutils [None req-f2193767-9fb7-43c1-a839-d9595c16ee0b tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1395.613578] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1395.613796] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1396.116887] env[65107]: INFO nova.compute.manager [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Detaching volume 4bde42ea-1ce0-45f7-9ea7-17c50ae3db26 [ 1396.147988] env[65107]: INFO nova.virt.block_device [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Attempting to driver detach volume 4bde42ea-1ce0-45f7-9ea7-17c50ae3db26 from mountpoint /dev/sdb [ 1396.148812] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1396.148812] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992917', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'name': 'volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'serial': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1396.149352] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa38a53-cf66-4290-805b-7cb4fe051b10 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.174891] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa67db5-e50e-4cdc-8975-92aa2fa231c2 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.182455] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6603e22d-f282-4534-9cfe-32dc6df74d3d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.205492] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41ada0d-77b5-4407-9dc2-d03d3fd69f8a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.221294] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The volume has not been displaced from its original location: [datastore1] volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26/volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1396.226382] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfiguring VM instance instance-0000007d to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1396.226668] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d086d693-6860-4698-bca0-1ef3c5de9108 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.245075] env[65107]: DEBUG oslo_vmware.api [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1396.245075] env[65107]: value = "task-5103952" [ 1396.245075] env[65107]: _type = "Task" [ 1396.245075] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.255356] env[65107]: DEBUG oslo_vmware.api [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103952, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.754862] env[65107]: DEBUG oslo_vmware.api [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103952, 'name': ReconfigVM_Task, 'duration_secs': 0.217821} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.755143] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfigured VM instance instance-0000007d to detach disk 2001 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1396.759874] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01a5ed6a-c53e-46cd-8219-82ebac448519 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.775952] env[65107]: DEBUG oslo_vmware.api [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1396.775952] env[65107]: value = "task-5103953" [ 1396.775952] env[65107]: _type = "Task" [ 1396.775952] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.784346] env[65107]: DEBUG oslo_vmware.api [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103953, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.285958] env[65107]: DEBUG oslo_vmware.api [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103953, 'name': ReconfigVM_Task, 'duration_secs': 0.143234} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.286355] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992917', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'name': 'volume-4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26', 'serial': '4bde42ea-1ce0-45f7-9ea7-17c50ae3db26'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1397.829030] env[65107]: DEBUG nova.objects.instance [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid f354b9e3-c426-48ca-8fe6-8bee94899ec8 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1398.144306] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "995c579b-9e18-4d3d-816d-0b48f352f938" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1398.144572] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1398.144742] env[65107]: DEBUG nova.compute.manager [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1398.145701] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75325cf7-af94-40fa-bf68-afc8ff81d1c7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.152526] env[65107]: DEBUG nova.compute.manager [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65107) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1398.153116] env[65107]: DEBUG nova.objects.instance [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'flavor' on Instance uuid 995c579b-9e18-4d3d-816d-0b48f352f938 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1398.839050] env[65107]: DEBUG oslo_concurrency.lockutils [None req-b9f2a1aa-3f45-4128-96b0-5c700b70877e tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1398.858027] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1398.858223] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1399.160493] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1399.160838] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba6fcc1c-8a5e-4749-aa33-90f2525bbdba {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.170010] env[65107]: DEBUG oslo_vmware.api [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1399.170010] env[65107]: value = "task-5103954" [ 1399.170010] env[65107]: _type = "Task" [ 1399.170010] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.178795] env[65107]: DEBUG oslo_vmware.api [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.361742] env[65107]: INFO nova.compute.manager [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Detaching volume bc475fc5-e5ae-48d2-97c4-33913a95981e [ 1399.396541] env[65107]: INFO nova.virt.block_device [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Attempting to driver detach volume bc475fc5-e5ae-48d2-97c4-33913a95981e from mountpoint /dev/sdc [ 1399.396803] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Volume detach. Driver type: vmdk {{(pid=65107) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1399.397012] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992918', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'name': 'volume-bc475fc5-e5ae-48d2-97c4-33913a95981e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'serial': 'bc475fc5-e5ae-48d2-97c4-33913a95981e'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1399.398038] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c3db2d-5809-49bb-acc6-fe0b3ed64585 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.421328] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b3c1d2-af69-4426-8246-ebafcf8501a6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.428619] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c977ec70-c77c-4e8d-ab16-bde6a11ec16c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.448943] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53504642-2242-442f-bff7-66a6ceecf954 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.464011] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] The volume has not been displaced from its original location: [datastore1] volume-bc475fc5-e5ae-48d2-97c4-33913a95981e/volume-bc475fc5-e5ae-48d2-97c4-33913a95981e.vmdk. No consolidation needed. {{(pid=65107) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1399.469358] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfiguring VM instance instance-0000007d to detach disk 2002 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1399.469670] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c02e8b6-e730-43a1-8d57-73ee2ef9b342 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.490217] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1399.490217] env[65107]: value = "task-5103955" [ 1399.490217] env[65107]: _type = "Task" [ 1399.490217] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.498353] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103955, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.679837] env[65107]: DEBUG oslo_vmware.api [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103954, 'name': PowerOffVM_Task, 'duration_secs': 0.174112} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1399.680214] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1399.680384] env[65107]: DEBUG nova.compute.manager [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1399.681185] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a045f4f6-576b-4905-a1c5-528f38f1fe64 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.003939] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103955, 'name': ReconfigVM_Task, 'duration_secs': 0.205008} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.004274] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Reconfigured VM instance instance-0000007d to detach disk 2002 {{(pid=65107) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1400.008878] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e34bb61-96a1-446d-a5b2-9759fa7c4997 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.024489] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1400.024489] env[65107]: value = "task-5103956" [ 1400.024489] env[65107]: _type = "Task" [ 1400.024489] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.033076] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103956, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.193221] env[65107]: DEBUG oslo_concurrency.lockutils [None req-ded910d7-cc49-4a48-9a17-b4e1dcd397eb tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.049s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1400.534851] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103956, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.548836] env[65107]: DEBUG nova.objects.instance [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'flavor' on Instance uuid 995c579b-9e18-4d3d-816d-0b48f352f938 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1401.034882] env[65107]: DEBUG oslo_vmware.api [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103956, 'name': ReconfigVM_Task, 'duration_secs': 0.786444} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1401.035319] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-992918', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'name': 'volume-bc475fc5-e5ae-48d2-97c4-33913a95981e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f354b9e3-c426-48ca-8fe6-8bee94899ec8', 'attached_at': '', 'detached_at': '', 'volume_id': 'bc475fc5-e5ae-48d2-97c4-33913a95981e', 'serial': 'bc475fc5-e5ae-48d2-97c4-33913a95981e'} {{(pid=65107) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1401.053805] env[65107]: DEBUG oslo_concurrency.lockutils [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1401.053985] env[65107]: DEBUG oslo_concurrency.lockutils [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1401.054191] env[65107]: DEBUG nova.network.neutron [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1401.054372] env[65107]: DEBUG nova.objects.instance [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'info_cache' on Instance uuid 995c579b-9e18-4d3d-816d-0b48f352f938 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1401.558011] env[65107]: DEBUG nova.objects.base [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Object Instance<995c579b-9e18-4d3d-816d-0b48f352f938> lazy-loaded attributes: flavor,info_cache {{(pid=65107) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1401.576703] env[65107]: DEBUG nova.objects.instance [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'flavor' on Instance uuid f354b9e3-c426-48ca-8fe6-8bee94899ec8 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1402.060386] env[65107]: WARNING neutronclient.v2_0.client [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1402.061116] env[65107]: WARNING openstack [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1402.061520] env[65107]: WARNING openstack [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1402.237104] env[65107]: WARNING openstack [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1402.237590] env[65107]: WARNING openstack [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1402.306192] env[65107]: WARNING neutronclient.v2_0.client [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1402.306933] env[65107]: WARNING openstack [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1402.307426] env[65107]: WARNING openstack [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1402.385786] env[65107]: DEBUG nova.network.neutron [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [{"id": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "address": "fa:16:3e:9e:0c:e8", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap187b4d9a-9a", "ovs_interfaceid": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1402.583644] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5b31fa65-0ba7-4b93-93ac-05f42d77be63 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.725s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1402.889044] env[65107]: DEBUG oslo_concurrency.lockutils [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1403.251218] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1403.251741] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1403.251741] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1403.251888] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1403.252042] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1403.254434] env[65107]: INFO nova.compute.manager [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Terminating instance [ 1403.758741] env[65107]: DEBUG nova.compute.manager [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1403.758968] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1403.759967] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260ccfcb-9d8c-4227-b0a7-6a723d7f0a1f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.769953] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1403.770209] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eacaaa09-240e-44a0-bab7-e43ae1067c38 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.777101] env[65107]: DEBUG oslo_vmware.api [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1403.777101] env[65107]: value = "task-5103957" [ 1403.777101] env[65107]: _type = "Task" [ 1403.777101] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.786187] env[65107]: DEBUG oslo_vmware.api [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.895263] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1403.895672] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b0f01c9-215d-4d27-9ad7-451b45439d11 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.903475] env[65107]: DEBUG oslo_vmware.api [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1403.903475] env[65107]: value = "task-5103958" [ 1403.903475] env[65107]: _type = "Task" [ 1403.903475] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1403.912791] env[65107]: DEBUG oslo_vmware.api [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103958, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.288096] env[65107]: DEBUG oslo_vmware.api [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103957, 'name': PowerOffVM_Task, 'duration_secs': 0.199858} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.288502] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1404.288595] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1404.288849] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-04f2332f-eb70-4ccb-8fa9-9f40096cc2c6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.356242] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1404.356435] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1404.356608] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Deleting the datastore file [datastore1] f354b9e3-c426-48ca-8fe6-8bee94899ec8 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1404.356892] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fca049d-6670-4177-8c59-b60aa0875230 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.364658] env[65107]: DEBUG oslo_vmware.api [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for the task: (returnval){ [ 1404.364658] env[65107]: value = "task-5103960" [ 1404.364658] env[65107]: _type = "Task" [ 1404.364658] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.373907] env[65107]: DEBUG oslo_vmware.api [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.413828] env[65107]: DEBUG oslo_vmware.api [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103958, 'name': PowerOnVM_Task, 'duration_secs': 0.373656} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.414125] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1404.414344] env[65107]: DEBUG nova.compute.manager [None req-519dfde4-8d5e-4274-baee-2b96343c4b0f tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1404.415227] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199bdeaa-c667-408a-a819-239f98539f44 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.875640] env[65107]: DEBUG oslo_vmware.api [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Task: {'id': task-5103960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231549} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.875816] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1404.875907] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1404.876068] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1404.876245] env[65107]: INFO nova.compute.manager [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1404.876511] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1404.876707] env[65107]: DEBUG nova.compute.manager [-] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1404.876804] env[65107]: DEBUG nova.network.neutron [-] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1404.877070] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1404.877651] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1404.877921] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1404.951230] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1405.359230] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f7346f-1e8d-4b7a-974c-7c08fd9d7996 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.366653] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Suspending the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1405.366987] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-f4abef69-1b03-46a3-a05b-6100335f3d4c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.377154] env[65107]: DEBUG oslo_vmware.api [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1405.377154] env[65107]: value = "task-5103961" [ 1405.377154] env[65107]: _type = "Task" [ 1405.377154] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.387424] env[65107]: DEBUG oslo_vmware.api [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103961, 'name': SuspendVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.420903] env[65107]: DEBUG nova.compute.manager [req-dd80c667-6465-47f0-bec3-e7fcfc6ee1fc req-9f8fb6b8-8b82-470f-8042-26d22871d5d4 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Received event network-vif-deleted-a900303a-2f4a-4b1f-b115-37e160d3834a {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1405.421053] env[65107]: INFO nova.compute.manager [req-dd80c667-6465-47f0-bec3-e7fcfc6ee1fc req-9f8fb6b8-8b82-470f-8042-26d22871d5d4 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Neutron deleted interface a900303a-2f4a-4b1f-b115-37e160d3834a; detaching it from the instance and deleting it from the info cache [ 1405.421238] env[65107]: DEBUG nova.network.neutron [req-dd80c667-6465-47f0-bec3-e7fcfc6ee1fc req-9f8fb6b8-8b82-470f-8042-26d22871d5d4 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1405.887321] env[65107]: DEBUG oslo_vmware.api [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103961, 'name': SuspendVM_Task} progress is 66%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.896842] env[65107]: DEBUG nova.network.neutron [-] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1405.924387] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c32333f2-43b4-4ccb-92fd-4dc44e0d4d01 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.935885] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f4a710-e304-4aa0-836e-1177627b044c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.966869] env[65107]: DEBUG nova.compute.manager [req-dd80c667-6465-47f0-bec3-e7fcfc6ee1fc req-9f8fb6b8-8b82-470f-8042-26d22871d5d4 service nova] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Detach interface failed, port_id=a900303a-2f4a-4b1f-b115-37e160d3834a, reason: Instance f354b9e3-c426-48ca-8fe6-8bee94899ec8 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1406.387207] env[65107]: DEBUG oslo_vmware.api [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103961, 'name': SuspendVM_Task, 'duration_secs': 0.617151} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.387597] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Suspended the VM {{(pid=65107) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1406.387715] env[65107]: DEBUG nova.compute.manager [None req-b6df32ae-6a94-4a06-ab94-d4c24fd8b2fa tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1406.388503] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72064de1-f0e8-49da-a28c-affdc6b4956f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.400064] env[65107]: INFO nova.compute.manager [-] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Took 1.52 seconds to deallocate network for instance. [ 1406.904963] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1406.905914] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1406.905914] env[65107]: DEBUG nova.objects.instance [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lazy-loading 'resources' on Instance uuid f354b9e3-c426-48ca-8fe6-8bee94899ec8 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1407.449491] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a149630-29d1-4b70-b066-f76abaa95e3b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.457096] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e06aa96-616b-4e42-878a-0d9299a6d552 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.488996] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cfc7fc-0aea-4faa-a37b-22f5c3ba1392 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.496675] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5964b733-d88e-43f8-80c4-ee4ab2290fb9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.510571] env[65107]: DEBUG nova.compute.provider_tree [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1407.749281] env[65107]: INFO nova.compute.manager [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Resuming [ 1407.750056] env[65107]: DEBUG nova.objects.instance [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'flavor' on Instance uuid 995c579b-9e18-4d3d-816d-0b48f352f938 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1408.014224] env[65107]: DEBUG nova.scheduler.client.report [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1408.520802] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1408.544168] env[65107]: INFO nova.scheduler.client.report [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Deleted allocations for instance f354b9e3-c426-48ca-8fe6-8bee94899ec8 [ 1409.052459] env[65107]: DEBUG oslo_concurrency.lockutils [None req-5412add6-94ab-4c02-a60a-145f18e94707 tempest-AttachVolumeTestJSON-843685348 tempest-AttachVolumeTestJSON-843685348-project-member] Lock "f354b9e3-c426-48ca-8fe6-8bee94899ec8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.801s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1409.260262] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1409.260486] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquired lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1409.260644] env[65107]: DEBUG nova.network.neutron [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1409.764467] env[65107]: WARNING neutronclient.v2_0.client [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1409.765250] env[65107]: WARNING openstack [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1409.765650] env[65107]: WARNING openstack [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1409.873634] env[65107]: WARNING openstack [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1409.874109] env[65107]: WARNING openstack [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1409.934445] env[65107]: WARNING neutronclient.v2_0.client [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1409.935153] env[65107]: WARNING openstack [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1409.935497] env[65107]: WARNING openstack [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1410.014914] env[65107]: DEBUG nova.network.neutron [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [{"id": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "address": "fa:16:3e:9e:0c:e8", "network": {"id": "9ee4166a-66dc-4bee-8dec-1f395aba1f8c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1486302518-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "139a38d3b7bd4d3a8861d2e1e6c7a2c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f9ffee1-f413-4f28-8bc4-3fb2cf299789", "external-id": "nsx-vlan-transportzone-599", "segmentation_id": 599, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap187b4d9a-9a", "ovs_interfaceid": "187b4d9a-9a89-4e19-8674-5bd91faf1d78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1410.517762] env[65107]: DEBUG oslo_concurrency.lockutils [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Releasing lock "refresh_cache-995c579b-9e18-4d3d-816d-0b48f352f938" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1410.518746] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30cfd43-29f3-430c-a082-7bc25150fc0a {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.525656] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Resuming the VM {{(pid=65107) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1410.525914] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54a3ad20-8384-4610-9574-15e606ee8e25 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.533326] env[65107]: DEBUG oslo_vmware.api [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1410.533326] env[65107]: value = "task-5103963" [ 1410.533326] env[65107]: _type = "Task" [ 1410.533326] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.541266] env[65107]: DEBUG oslo_vmware.api [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.044493] env[65107]: DEBUG oslo_vmware.api [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103963, 'name': PowerOnVM_Task, 'duration_secs': 0.504734} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.044937] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Resumed the VM {{(pid=65107) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1411.044937] env[65107]: DEBUG nova.compute.manager [None req-a60a61e5-5b4f-489c-95d5-8b4a9bceaad4 tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1411.046036] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add94135-eb61-45c5-a81d-9c0e36608ed7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.524462] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "995c579b-9e18-4d3d-816d-0b48f352f938" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1412.524906] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1412.525016] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "995c579b-9e18-4d3d-816d-0b48f352f938-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1412.525345] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1412.525532] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1412.527809] env[65107]: INFO nova.compute.manager [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Terminating instance [ 1413.032184] env[65107]: DEBUG nova.compute.manager [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1413.032453] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1413.033798] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6662f1f3-1ba2-44ae-9b1e-d11a52fbfa76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.041544] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1413.041743] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-acf17e85-94e9-4f59-bbe4-f777ad57c2e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.048167] env[65107]: DEBUG oslo_vmware.api [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1413.048167] env[65107]: value = "task-5103965" [ 1413.048167] env[65107]: _type = "Task" [ 1413.048167] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1413.059349] env[65107]: DEBUG oslo_vmware.api [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1413.559765] env[65107]: DEBUG oslo_vmware.api [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103965, 'name': PowerOffVM_Task, 'duration_secs': 0.208176} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1413.560256] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1413.560552] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1413.560855] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc7cadb1-3968-4a99-8383-770f1484bdef {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.633491] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1413.633945] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Deleting contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1413.634252] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleting the datastore file [datastore2] 995c579b-9e18-4d3d-816d-0b48f352f938 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1413.634599] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-69392ac4-890a-41ab-9e8f-0f36292ffb67 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.648371] env[65107]: DEBUG oslo_vmware.api [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for the task: (returnval){ [ 1413.648371] env[65107]: value = "task-5103967" [ 1413.648371] env[65107]: _type = "Task" [ 1413.648371] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1413.657034] env[65107]: DEBUG oslo_vmware.api [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.158859] env[65107]: DEBUG oslo_vmware.api [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Task: {'id': task-5103967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157691} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.159077] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1414.159251] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Deleted contents of the VM from datastore datastore2 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1414.159426] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1414.159603] env[65107]: INFO nova.compute.manager [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1414.159853] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1414.160061] env[65107]: DEBUG nova.compute.manager [-] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1414.160171] env[65107]: DEBUG nova.network.neutron [-] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1414.160400] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1414.160918] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1414.161188] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1414.227948] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1414.675716] env[65107]: DEBUG nova.compute.manager [req-730bafbc-81df-40a9-8d27-1ede5f91afe7 req-8ac39d47-5090-4f7c-889e-0fcef1b0bd8b service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Received event network-vif-deleted-187b4d9a-9a89-4e19-8674-5bd91faf1d78 {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1414.675986] env[65107]: INFO nova.compute.manager [req-730bafbc-81df-40a9-8d27-1ede5f91afe7 req-8ac39d47-5090-4f7c-889e-0fcef1b0bd8b service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Neutron deleted interface 187b4d9a-9a89-4e19-8674-5bd91faf1d78; detaching it from the instance and deleting it from the info cache [ 1414.676167] env[65107]: DEBUG nova.network.neutron [req-730bafbc-81df-40a9-8d27-1ede5f91afe7 req-8ac39d47-5090-4f7c-889e-0fcef1b0bd8b service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1415.154095] env[65107]: DEBUG nova.network.neutron [-] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1415.178365] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-807a7798-97e4-4ca4-a744-32ed0695d1c0 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.190057] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1cdbe6-8dd5-439a-9af0-d78960e8675f {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.217112] env[65107]: DEBUG nova.compute.manager [req-730bafbc-81df-40a9-8d27-1ede5f91afe7 req-8ac39d47-5090-4f7c-889e-0fcef1b0bd8b service nova] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Detach interface failed, port_id=187b4d9a-9a89-4e19-8674-5bd91faf1d78, reason: Instance 995c579b-9e18-4d3d-816d-0b48f352f938 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1415.657759] env[65107]: INFO nova.compute.manager [-] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Took 1.50 seconds to deallocate network for instance. [ 1416.164401] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1416.164796] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1416.164844] env[65107]: DEBUG nova.objects.instance [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lazy-loading 'resources' on Instance uuid 995c579b-9e18-4d3d-816d-0b48f352f938 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1416.699648] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3df47f-5f5a-4204-9092-43eeecd6d6dc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.707182] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85eb27c1-5111-4214-901d-3405cbc02639 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.739256] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdc152e-6027-45ae-bc8c-a0f972744943 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.746957] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3432abe6-3665-46f1-909e-d88024e092da {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.761246] env[65107]: DEBUG nova.compute.provider_tree [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1417.264847] env[65107]: DEBUG nova.scheduler.client.report [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1417.770718] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.606s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1417.795642] env[65107]: INFO nova.scheduler.client.report [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Deleted allocations for instance 995c579b-9e18-4d3d-816d-0b48f352f938 [ 1418.303154] env[65107]: DEBUG oslo_concurrency.lockutils [None req-d6eb15f2-7c47-4a21-aa01-d27def0f58cd tempest-ServerActionsTestJSON-273915797 tempest-ServerActionsTestJSON-273915797-project-member] Lock "995c579b-9e18-4d3d-816d-0b48f352f938" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.778s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1423.988837] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1423.989213] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1424.492717] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Starting instance... {{(pid=65107) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1425.014559] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1425.014972] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1425.016959] env[65107]: INFO nova.compute.claims [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1426.058707] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041313f0-ecfc-4429-8c55-6e03c7f40529 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.066705] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbcc920-dc5c-4fc5-8905-57219bd15b6c {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.098123] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53612cc-2cbf-46cc-9d20-8769d36c8cc9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.105830] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae70b007-93fb-4c84-98a9-7122ec6ea003 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1426.119178] env[65107]: DEBUG nova.compute.provider_tree [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1426.622622] env[65107]: DEBUG nova.scheduler.client.report [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1427.127725] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.113s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1427.128336] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Start building networks asynchronously for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1427.633841] env[65107]: DEBUG nova.compute.utils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Using /dev/sd instead of None {{(pid=65107) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1427.635965] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Allocating IP information in the background. {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1427.636229] env[65107]: DEBUG nova.network.neutron [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] allocate_for_instance() {{(pid=65107) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1427.636599] env[65107]: WARNING neutronclient.v2_0.client [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1427.636978] env[65107]: WARNING neutronclient.v2_0.client [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1427.637676] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1427.638132] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1427.689764] env[65107]: DEBUG nova.policy [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5946042e9faa4f96b48b7183963e3a61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21cf13096bc740149eb899062b6f5241', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65107) authorize /opt/stack/nova/nova/policy.py:192}} [ 1427.966649] env[65107]: DEBUG nova.network.neutron [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Successfully created port: 3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1428.146200] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Start building block device mappings for instance. {{(pid=65107) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1429.156073] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Start spawning the instance on the hypervisor. {{(pid=65107) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1429.182588] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T13:19:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T13:18:40Z,direct_url=,disk_format='vmdk',id=e439aeac-2298-462c-be63-8218195135cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5b452d0d7ba0461d97515e1de6eeeb7d',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T13:18:40Z,virtual_size=,visibility=), allow threads: False {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1429.182884] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Flavor limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1429.183077] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Image limits 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1429.183273] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Flavor pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1429.183421] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Image pref 0:0:0 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1429.183569] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65107) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1429.183779] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1429.183939] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1429.184119] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Got 1 possible topologies {{(pid=65107) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1429.184298] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1429.184501] env[65107]: DEBUG nova.virt.hardware [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65107) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1429.185401] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c633676-7fae-44d0-b297-edd153340f84 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.194648] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16eb2a54-9313-4a19-b2fd-ca39290c3edc {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.357362] env[65107]: DEBUG nova.compute.manager [req-f2aa0d33-ff65-485a-af28-ff848d1063d9 req-1aef2e81-1538-4336-8f16-02c892020768 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Received event network-vif-plugged-3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1429.357729] env[65107]: DEBUG oslo_concurrency.lockutils [req-f2aa0d33-ff65-485a-af28-ff848d1063d9 req-1aef2e81-1538-4336-8f16-02c892020768 service nova] Acquiring lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1429.357835] env[65107]: DEBUG oslo_concurrency.lockutils [req-f2aa0d33-ff65-485a-af28-ff848d1063d9 req-1aef2e81-1538-4336-8f16-02c892020768 service nova] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1429.357981] env[65107]: DEBUG oslo_concurrency.lockutils [req-f2aa0d33-ff65-485a-af28-ff848d1063d9 req-1aef2e81-1538-4336-8f16-02c892020768 service nova] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1429.358793] env[65107]: DEBUG nova.compute.manager [req-f2aa0d33-ff65-485a-af28-ff848d1063d9 req-1aef2e81-1538-4336-8f16-02c892020768 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] No waiting events found dispatching network-vif-plugged-3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1429.358793] env[65107]: WARNING nova.compute.manager [req-f2aa0d33-ff65-485a-af28-ff848d1063d9 req-1aef2e81-1538-4336-8f16-02c892020768 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Received unexpected event network-vif-plugged-3161afbe-814d-44e0-8611-720927ba01eb for instance with vm_state building and task_state spawning. [ 1429.455245] env[65107]: DEBUG nova.network.neutron [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Successfully updated port: 3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1429.958375] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "refresh_cache-2b007ad8-8f5d-49c1-9388-4aeadb09d190" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1429.958587] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquired lock "refresh_cache-2b007ad8-8f5d-49c1-9388-4aeadb09d190" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1429.958760] env[65107]: DEBUG nova.network.neutron [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Building network info cache for instance {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2104}} [ 1430.464604] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1430.465066] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1430.503307] env[65107]: DEBUG nova.network.neutron [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Instance cache missing network info. {{(pid=65107) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3433}} [ 1430.523403] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1430.523861] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1430.587521] env[65107]: WARNING neutronclient.v2_0.client [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1430.588286] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1430.588633] env[65107]: WARNING openstack [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1430.667498] env[65107]: DEBUG nova.network.neutron [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Updating instance_info_cache with network_info: [{"id": "3161afbe-814d-44e0-8611-720927ba01eb", "address": "fa:16:3e:91:e7:37", "network": {"id": "d866e3c8-db8a-4afd-817b-db496f45fca4", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-588911655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21cf13096bc740149eb899062b6f5241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3161afbe-81", "ovs_interfaceid": "3161afbe-814d-44e0-8611-720927ba01eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1431.170400] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Releasing lock "refresh_cache-2b007ad8-8f5d-49c1-9388-4aeadb09d190" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1431.170780] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Instance network_info: |[{"id": "3161afbe-814d-44e0-8611-720927ba01eb", "address": "fa:16:3e:91:e7:37", "network": {"id": "d866e3c8-db8a-4afd-817b-db496f45fca4", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-588911655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21cf13096bc740149eb899062b6f5241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3161afbe-81", "ovs_interfaceid": "3161afbe-814d-44e0-8611-720927ba01eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65107) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1431.171286] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:e7:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3161afbe-814d-44e0-8611-720927ba01eb', 'vif_model': 'vmxnet3'}] {{(pid=65107) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1431.178713] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Creating folder: Project (21cf13096bc740149eb899062b6f5241). Parent ref: group-v992574. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1431.178982] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16ec1497-4d44-411b-b2b5-47e606606af5 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.190761] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Created folder: Project (21cf13096bc740149eb899062b6f5241) in parent group-v992574. [ 1431.190976] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Creating folder: Instances. Parent ref: group-v992919. {{(pid=65107) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1431.191234] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e986be3-66f1-45ef-9ba8-8ba5fcfb2353 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.201915] env[65107]: INFO nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Created folder: Instances in parent group-v992919. [ 1431.202178] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1431.202380] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Creating VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1431.202590] env[65107]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c736f717-5157-4674-8b23-52da1235c94d {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.222919] env[65107]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1431.222919] env[65107]: value = "task-5103971" [ 1431.222919] env[65107]: _type = "Task" [ 1431.222919] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.231153] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103971, 'name': CreateVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.387899] env[65107]: DEBUG nova.compute.manager [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Received event network-changed-3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1431.388158] env[65107]: DEBUG nova.compute.manager [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Refreshing instance network info cache due to event network-changed-3161afbe-814d-44e0-8611-720927ba01eb. {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1431.388443] env[65107]: DEBUG oslo_concurrency.lockutils [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Acquiring lock "refresh_cache-2b007ad8-8f5d-49c1-9388-4aeadb09d190" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1431.388652] env[65107]: DEBUG oslo_concurrency.lockutils [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Acquired lock "refresh_cache-2b007ad8-8f5d-49c1-9388-4aeadb09d190" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1431.388897] env[65107]: DEBUG nova.network.neutron [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Refreshing network info cache for port 3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2101}} [ 1431.732625] env[65107]: DEBUG oslo_vmware.api [-] Task: {'id': task-5103971, 'name': CreateVM_Task, 'duration_secs': 0.322469} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.733025] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Created VM on the ESX host {{(pid=65107) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1431.733330] env[65107]: WARNING neutronclient.v2_0.client [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1431.733696] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1431.733852] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1431.734263] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1431.734546] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2767580-641b-4116-bc45-be2255748c18 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.739580] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1431.739580] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f77cb5-12be-669a-0e6c-b03554e1cdde" [ 1431.739580] env[65107]: _type = "Task" [ 1431.739580] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.747665] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f77cb5-12be-669a-0e6c-b03554e1cdde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.891714] env[65107]: WARNING neutronclient.v2_0.client [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1431.892453] env[65107]: WARNING openstack [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1431.892797] env[65107]: WARNING openstack [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1431.985829] env[65107]: WARNING openstack [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1431.985829] env[65107]: WARNING openstack [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1432.045584] env[65107]: WARNING neutronclient.v2_0.client [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1432.046370] env[65107]: WARNING openstack [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1432.046646] env[65107]: WARNING openstack [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1432.127122] env[65107]: DEBUG nova.network.neutron [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Updated VIF entry in instance network info cache for port 3161afbe-814d-44e0-8611-720927ba01eb. {{(pid=65107) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3593}} [ 1432.127594] env[65107]: DEBUG nova.network.neutron [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Updating instance_info_cache with network_info: [{"id": "3161afbe-814d-44e0-8611-720927ba01eb", "address": "fa:16:3e:91:e7:37", "network": {"id": "d866e3c8-db8a-4afd-817b-db496f45fca4", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-588911655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "21cf13096bc740149eb899062b6f5241", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3161afbe-81", "ovs_interfaceid": "3161afbe-814d-44e0-8611-720927ba01eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1432.250305] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52f77cb5-12be-669a-0e6c-b03554e1cdde, 'name': SearchDatastore_Task, 'duration_secs': 0.011635} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.250533] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1432.250770] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Processing image e439aeac-2298-462c-be63-8218195135cf {{(pid=65107) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1432.251016] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1432.251164] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1432.251339] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1432.251596] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb907458-94eb-4f03-a9e0-a9d5e05f2195 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.261049] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65107) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1432.261341] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65107) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1432.262445] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd5f0b30-505d-47fd-a1cc-532a677f3372 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.269607] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1432.269607] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521f58f7-6ce7-be01-5d12-dab7c19a2d50" [ 1432.269607] env[65107]: _type = "Task" [ 1432.269607] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.283967] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]521f58f7-6ce7-be01-5d12-dab7c19a2d50, 'name': SearchDatastore_Task, 'duration_secs': 0.00988} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.285261] env[65107]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daa893f6-8b79-4b78-a08a-c980cc7d153e {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.292790] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1432.292790] env[65107]: value = "session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7abcf-23f3-0763-f7cc-a9784f902d52" [ 1432.292790] env[65107]: _type = "Task" [ 1432.292790] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.306549] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': session[528a712f-48b1-e08f-fcdd-3eca1a0ef49b]52c7abcf-23f3-0763-f7cc-a9784f902d52, 'name': SearchDatastore_Task, 'duration_secs': 0.009517} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.306946] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1432.307372] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2b007ad8-8f5d-49c1-9388-4aeadb09d190/2b007ad8-8f5d-49c1-9388-4aeadb09d190.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1432.307746] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61a59df1-cf0e-4668-909d-818335de00d8 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.317020] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1432.317020] env[65107]: value = "task-5103972" [ 1432.317020] env[65107]: _type = "Task" [ 1432.317020] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.330078] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1432.630732] env[65107]: DEBUG oslo_concurrency.lockutils [req-49e6fc87-ccfc-4371-9cec-ed9d0ef9f47a req-4785e5e8-1726-4d84-9484-1c4d1a1e6b11 service nova] Releasing lock "refresh_cache-2b007ad8-8f5d-49c1-9388-4aeadb09d190" {{(pid=65107) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1432.828832] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451868} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.829336] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e439aeac-2298-462c-be63-8218195135cf/e439aeac-2298-462c-be63-8218195135cf.vmdk to [datastore1] 2b007ad8-8f5d-49c1-9388-4aeadb09d190/2b007ad8-8f5d-49c1-9388-4aeadb09d190.vmdk {{(pid=65107) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1432.829651] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Extending root virtual disk to 1048576 {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1432.830065] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-89ae7360-9fa8-4828-972d-bbd5fd528902 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.838466] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1432.838466] env[65107]: value = "task-5103973" [ 1432.838466] env[65107]: _type = "Task" [ 1432.838466] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.852609] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.349114] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066882} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.349364] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Extended root virtual disk {{(pid=65107) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1433.350154] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735976ba-44c1-4be4-9ac4-3637f48d2c76 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.371965] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Reconfiguring VM instance instance-0000007f to attach disk [datastore1] 2b007ad8-8f5d-49c1-9388-4aeadb09d190/2b007ad8-8f5d-49c1-9388-4aeadb09d190.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1433.372272] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c7f4a8a-677f-4df7-8861-a08c6e1f3d20 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.392231] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1433.392231] env[65107]: value = "task-5103974" [ 1433.392231] env[65107]: _type = "Task" [ 1433.392231] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.400559] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.499919] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1433.903246] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103974, 'name': ReconfigVM_Task, 'duration_secs': 0.316674} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.903668] env[65107]: DEBUG nova.virt.vmwareapi.volumeops [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Reconfigured VM instance instance-0000007f to attach disk [datastore1] 2b007ad8-8f5d-49c1-9388-4aeadb09d190/2b007ad8-8f5d-49c1-9388-4aeadb09d190.vmdk or device None with type sparse {{(pid=65107) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1433.904210] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-017536cc-3840-4b73-9e21-4dbedc951465 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.912039] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1433.912039] env[65107]: value = "task-5103975" [ 1433.912039] env[65107]: _type = "Task" [ 1433.912039] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.922081] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103975, 'name': Rename_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.422530] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103975, 'name': Rename_Task, 'duration_secs': 0.146484} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.422823] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Powering on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1434.423108] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e9de3b6-3e1d-461c-bac9-414ab20907eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.430041] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1434.430041] env[65107]: value = "task-5103976" [ 1434.430041] env[65107]: _type = "Task" [ 1434.430041] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1434.444249] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.939915] env[65107]: DEBUG oslo_vmware.api [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103976, 'name': PowerOnVM_Task, 'duration_secs': 0.451306} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.940394] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Powered on the VM {{(pid=65107) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1434.940394] env[65107]: INFO nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Took 5.78 seconds to spawn the instance on the hypervisor. [ 1434.940568] env[65107]: DEBUG nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Checking state {{(pid=65107) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1434.941337] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e672a1-71ff-4f77-98fc-10a9c022d266 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.459680] env[65107]: INFO nova.compute.manager [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Took 10.46 seconds to build instance. [ 1435.963216] env[65107]: DEBUG oslo_concurrency.lockutils [None req-9381967d-571b-44c4-905b-cc3c7d6bc155 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.974s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1436.057107] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1436.057439] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1436.057684] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1436.057943] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1436.058208] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1436.060446] env[65107]: INFO nova.compute.manager [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Terminating instance [ 1436.500033] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1436.500228] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65107) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1436.565380] env[65107]: DEBUG nova.compute.manager [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Start destroying the instance on the hypervisor. {{(pid=65107) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1436.565665] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Destroying instance {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1436.567067] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4357fe6-3708-4cb2-918b-29849069a4eb {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.574777] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Powering off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1436.575025] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c0d1890-3ade-4e55-a4ff-8da328a03794 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.581410] env[65107]: DEBUG oslo_vmware.api [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1436.581410] env[65107]: value = "task-5103977" [ 1436.581410] env[65107]: _type = "Task" [ 1436.581410] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1436.589545] env[65107]: DEBUG oslo_vmware.api [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103977, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1437.091424] env[65107]: DEBUG oslo_vmware.api [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103977, 'name': PowerOffVM_Task, 'duration_secs': 0.193698} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1437.091821] env[65107]: DEBUG nova.virt.vmwareapi.vm_util [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Powered off the VM {{(pid=65107) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1437.091916] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Unregistering the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1437.092196] env[65107]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-beb139f2-f441-4789-8991-05c4ccb5d6a3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.167614] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Unregistered the VM {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1437.167827] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Deleting contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1437.168024] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Deleting the datastore file [datastore1] 2b007ad8-8f5d-49c1-9388-4aeadb09d190 {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1437.168322] env[65107]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1211a104-7cf3-4f3a-a54d-317af4107e27 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1437.176991] env[65107]: DEBUG oslo_vmware.api [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for the task: (returnval){ [ 1437.176991] env[65107]: value = "task-5103979" [ 1437.176991] env[65107]: _type = "Task" [ 1437.176991] env[65107]: } to complete. {{(pid=65107) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1437.185819] env[65107]: DEBUG oslo_vmware.api [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103979, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1437.495865] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.499656] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.687216] env[65107]: DEBUG oslo_vmware.api [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Task: {'id': task-5103979, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136863} completed successfully. {{(pid=65107) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1437.687509] env[65107]: DEBUG nova.virt.vmwareapi.ds_util [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Deleted the datastore file {{(pid=65107) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1437.688034] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Deleted contents of the VM from datastore datastore1 {{(pid=65107) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1437.688034] env[65107]: DEBUG nova.virt.vmwareapi.vmops [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Instance destroyed {{(pid=65107) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1437.688281] env[65107]: INFO nova.compute.manager [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1437.688593] env[65107]: DEBUG oslo.service.backend._common.loopingcall [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65107) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1437.688808] env[65107]: DEBUG nova.compute.manager [-] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Deallocating network for instance {{(pid=65107) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1437.688907] env[65107]: DEBUG nova.network.neutron [-] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] deallocate_for_instance() {{(pid=65107) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1894}} [ 1437.689172] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1437.689721] env[65107]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1437.689977] env[65107]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1437.726211] env[65107]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1437.974874] env[65107]: DEBUG nova.compute.manager [req-2358cead-8591-4342-bb76-b515c3a119c5 req-b8a99942-cc3b-4134-a230-37517d5c378a service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Received event network-vif-deleted-3161afbe-814d-44e0-8611-720927ba01eb {{(pid=65107) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1437.975296] env[65107]: INFO nova.compute.manager [req-2358cead-8591-4342-bb76-b515c3a119c5 req-b8a99942-cc3b-4134-a230-37517d5c378a service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Neutron deleted interface 3161afbe-814d-44e0-8611-720927ba01eb; detaching it from the instance and deleting it from the info cache [ 1437.975296] env[65107]: DEBUG nova.network.neutron [req-2358cead-8591-4342-bb76-b515c3a119c5 req-b8a99942-cc3b-4134-a230-37517d5c378a service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1438.455145] env[65107]: DEBUG nova.network.neutron [-] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Updating instance_info_cache with network_info: [] {{(pid=65107) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1438.477652] env[65107]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-460f43a6-a055-422a-8301-472ae48822d6 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.488739] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c728220-0ef1-4a55-8230-bdc1f6f972a9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.499756] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1438.500448] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1438.500616] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Cleaning up deleted instances with incomplete migration {{(pid=65107) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 1438.515841] env[65107]: DEBUG nova.compute.manager [req-2358cead-8591-4342-bb76-b515c3a119c5 req-b8a99942-cc3b-4134-a230-37517d5c378a service nova] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Detach interface failed, port_id=3161afbe-814d-44e0-8611-720927ba01eb, reason: Instance 2b007ad8-8f5d-49c1-9388-4aeadb09d190 could not be found. {{(pid=65107) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1438.957762] env[65107]: INFO nova.compute.manager [-] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Took 1.27 seconds to deallocate network for instance. [ 1439.464834] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1439.465265] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1439.465421] env[65107]: DEBUG nova.objects.instance [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lazy-loading 'resources' on Instance uuid 2b007ad8-8f5d-49c1-9388-4aeadb09d190 {{(pid=65107) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1440.002954] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.066034] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fe869f-cb79-4b77-80f7-f41815955ef1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.074321] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf27acfd-5f77-408d-8f32-6b1ebe6522cd {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.106277] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e7af753-fe64-4019-8e32-b50e9bdbacd9 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.114701] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7f1af2-7501-4ea0-9834-ebb4f7e42869 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.129251] env[65107]: DEBUG nova.compute.provider_tree [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1440.500159] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.632650] env[65107]: DEBUG nova.scheduler.client.report [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1441.137978] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1441.159888] env[65107]: INFO nova.scheduler.client.report [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Deleted allocations for instance 2b007ad8-8f5d-49c1-9388-4aeadb09d190 [ 1441.668490] env[65107]: DEBUG oslo_concurrency.lockutils [None req-98cca2d6-8141-4552-afee-c2430cd89ce9 tempest-ServerAddressesTestJSON-2115740622 tempest-ServerAddressesTestJSON-2115740622-project-member] Lock "2b007ad8-8f5d-49c1-9388-4aeadb09d190" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.611s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1443.500672] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.500935] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Cleaning up deleted instances {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 1444.011240] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] There are 27 instances to clean {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 1444.011414] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 2b007ad8-8f5d-49c1-9388-4aeadb09d190] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1444.514400] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 995c579b-9e18-4d3d-816d-0b48f352f938] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1445.017937] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f354b9e3-c426-48ca-8fe6-8bee94899ec8] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1445.521153] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: f0a6cfb0-83ba-47ae-afb2-3aefdb83a4c6] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1446.024292] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 84f88238-6b21-4530-8a08-72c4049f9150] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1446.527638] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: e72d0cb3-027c-49af-aa23-b791ae2f5fcf] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1447.031249] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 2c80c51e-dff3-4e56-bc13-32b792f67d97] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1447.535018] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 10c2fd64-1f3b-41eb-a48c-34d4a892e2ad] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1448.038693] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 4a6894d7-233a-4c5c-8ded-5b80b9beadc7] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1448.542332] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 7e72891b-4c20-4889-80c9-8bcb15373a2c] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1449.046520] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 751ebb31-1a27-4630-9213-4312a6fa9298] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1449.550639] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d52a955c-bc3c-41f2-b3ab-db9da613279e] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1450.054345] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 4f93a52c-9c40-47e5-a34f-b19aea34d7a9] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1450.557703] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 0161c1ba-8aff-4f70-a407-caa991e43f15] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1451.061211] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: bcced66b-1464-4834-892b-3d0e652da31a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1451.564732] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3e0d9d2c-ed27-4790-bd09-19e6dcc23ea3] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1452.068294] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 475851f6-24a2-4f0e-9567-d2a23947a5e3] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1452.571841] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 93a1b356-b6d5-435b-b30f-ea5ade031b7e] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1453.075212] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 1f3292b7-f6d7-43ca-9caf-8c55794e8311] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1453.578618] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: bef91b92-add1-4df9-bc24-dab15ce04338] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1454.081882] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: d7b4bc1e-c94e-4654-9345-2a8aa945a896] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1454.585871] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 0599df85-ad33-4bd5-b234-611db1ef73f0] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1455.089696] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 709eaf30-f1a6-49c2-a0fd-a47eee8aef2d] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1455.593375] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 23b4fef2-7cdb-4cbe-9d21-2a546ccdf2e3] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1456.096680] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 2c45d232-a948-4ad7-80a0-589c2dc91ec1] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1456.600528] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: 3df65e64-0c45-4707-960e-8f1767e2d011] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1457.104286] env[65107]: DEBUG nova.compute.manager [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] [instance: cd481deb-b12c-47ab-9ab1-0c4bdbb2024a] Instance has had 0 of 5 cleanup attempts {{(pid=65107) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1457.608499] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.111574] env[65107]: DEBUG oslo_service.periodic_task [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65107) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.615112] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1459.615373] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1459.615544] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1459.615700] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65107) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1459.616650] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4929b06c-413d-4dd3-855c-6848d9ebce55 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.625391] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f2ba87-821d-487b-9887-ce38a9c40a6b {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.639285] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66541447-7579-44be-bf99-2091f4bbb3e7 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.646088] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bc31ec-e23c-4519-aa12-7f422ed46cf1 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1459.677721] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180508MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65107) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1459.677900] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1459.678086] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1460.699998] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1460.700314] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65107) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1460.714709] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cf10e2-c750-4c68-9e42-fd39cce7c887 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.723240] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a844694f-e435-4f2a-abe2-c6f599e6b513 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.754201] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42650670-a669-4bb1-ab71-1784d4282c95 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.762226] env[65107]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fcde16-ae99-4634-96b9-37b206059aa3 {{(pid=65107) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1460.777483] env[65107]: DEBUG nova.compute.provider_tree [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed in ProviderTree for provider: 8adccf8d-6d73-43c0-ad15-ee89967ab8cd {{(pid=65107) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1461.281068] env[65107]: DEBUG nova.scheduler.client.report [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Inventory has not changed for provider 8adccf8d-6d73-43c0-ad15-ee89967ab8cd based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65107) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1461.786208] env[65107]: DEBUG nova.compute.resource_tracker [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65107) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1461.786618] env[65107]: DEBUG oslo_concurrency.lockutils [None req-65d66000-665d-4b00-a763-f4c5f10f5460 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.108s {{(pid=65107) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}